[ 548.995727] env[66641]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=66641) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 548.996109] env[66641]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=66641) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 548.996266] env[66641]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=66641) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 548.996611] env[66641]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 549.123707] env[66641]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=66641) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:349}} [ 549.133933] env[66641]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=66641) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:372}} [ 549.183436] env[66641]: INFO oslo_service.periodic_task [-] Skipping periodic task _heal_instance_info_cache because its interval is negative [ 549.869032] env[66641]: INFO nova.virt.driver [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 549.942108] env[66641]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 549.942302] env[66641]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 549.942353] env[66641]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=66641) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 553.185329] env[66641]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-56eab7f9-e9eb-450a-bb82-fcba3712e88b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.201562] env[66641]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=66641) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 553.201774] env[66641]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-c70109f9-2a03-448e-80ea-86f6b1abcbde {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.230366] env[66641]: INFO oslo_vmware.api [-] Successfully established new session; session ID is dde9f. [ 553.230528] env[66641]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.288s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 553.231128] env[66641]: INFO nova.virt.vmwareapi.driver [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] VMware vCenter version: 7.0.3 [ 553.234621] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bc7913-c68e-40cf-8531-9bf9f02214f8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.253479] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c21c19-f9c0-4286-8023-64c97c61e46a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.260516] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c322fc55-a968-485f-92b3-a5571462c0c5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.267928] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304c29c1-3b30-4f0d-8371-483eaed33c12 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.281716] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d17b7e-fed1-473e-bb6c-637d8336827a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.288864] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ded03bd-9107-4672-9be1-dc77961ee417 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.319770] env[66641]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-4af59e27-78db-4ab6-8073-2a61c67a50d0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.325943] env[66641]: DEBUG nova.virt.vmwareapi.driver [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] Extension org.openstack.compute already exists. {{(pid=66641) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:228}} [ 553.328612] env[66641]: INFO nova.compute.provider_config [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 553.832472] env[66641]: DEBUG nova.context [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),a058062d-a8d0-4196-82de-41885581fe34(cell1) {{(pid=66641) load_cells /opt/stack/nova/nova/context.py:472}} [ 553.832773] env[66641]: INFO nova.utils [None req-2a2e6c35-2915-456c-92ce-4372a46541d9 None None] The cell worker thread pool MainProcess.cell_worker is initialized [ 553.834777] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 553.834994] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 553.835685] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 553.836142] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Acquiring lock "a058062d-a8d0-4196-82de-41885581fe34" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 553.836323] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Lock "a058062d-a8d0-4196-82de-41885581fe34" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 553.837430] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Lock "a058062d-a8d0-4196-82de-41885581fe34" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 553.860349] env[66641]: INFO dbcounter [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Registered counter for database nova_cell0 [ 553.868727] env[66641]: INFO dbcounter [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Registered counter for database nova_cell1 [ 553.872105] env[66641]: DEBUG oslo_db.sqlalchemy.engines [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=66641) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 553.872467] env[66641]: DEBUG oslo_db.sqlalchemy.engines [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=66641) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:397}} [ 553.877400] env[66641]: ERROR nova.db.main.api [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 553.877400] env[66641]: func(*args, **kwargs) [ 553.877400] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 553.877400] env[66641]: self.work.run() [ 553.877400] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 553.877400] env[66641]: result = self.fn(*self.args, **self.kwargs) [ 553.877400] env[66641]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 553.877400] env[66641]: return func(*args, **kwargs) [ 553.877400] env[66641]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 553.877400] env[66641]: result = fn(*args, **kwargs) [ 553.877400] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 553.877400] env[66641]: return f(*args, **kwargs) [ 553.877400] env[66641]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 553.877400] env[66641]: return db.service_get_minimum_version(context, binaries) [ 553.877400] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 553.877400] env[66641]: _check_db_access() [ 553.877400] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 553.877400] env[66641]: stacktrace = ''.join(traceback.format_stack()) [ 553.877400] env[66641]: [ 553.878276] env[66641]: ERROR nova.db.main.api [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenpool.py", line 87, in _spawn_n_impl [ 553.878276] env[66641]: func(*args, **kwargs) [ 553.878276] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_green.py", line 69, in __call__ [ 553.878276] env[66641]: self.work.run() [ 553.878276] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/futurist/_utils.py", line 45, in run [ 553.878276] env[66641]: result = self.fn(*self.args, **self.kwargs) [ 553.878276] env[66641]: File "/opt/stack/nova/nova/utils.py", line 585, in context_wrapper [ 553.878276] env[66641]: return func(*args, **kwargs) [ 553.878276] env[66641]: File "/opt/stack/nova/nova/context.py", line 420, in gather_result [ 553.878276] env[66641]: result = fn(*args, **kwargs) [ 553.878276] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 553.878276] env[66641]: return f(*args, **kwargs) [ 553.878276] env[66641]: File "/opt/stack/nova/nova/objects/service.py", line 568, in _db_service_get_minimum_version [ 553.878276] env[66641]: return db.service_get_minimum_version(context, binaries) [ 553.878276] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 553.878276] env[66641]: _check_db_access() [ 553.878276] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 553.878276] env[66641]: stacktrace = ''.join(traceback.format_stack()) [ 553.878276] env[66641]: [ 553.878753] env[66641]: WARNING nova.objects.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 553.878869] env[66641]: WARNING nova.objects.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Failed to get minimum service version for cell a058062d-a8d0-4196-82de-41885581fe34 [ 553.879315] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Acquiring lock "singleton_lock" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.879480] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Acquired lock "singleton_lock" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 553.879712] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Releasing lock "singleton_lock" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 553.880039] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Full set of CONF: {{(pid=66641) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_eventlet/service.py:275}} [ 553.880181] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ******************************************************************************** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2804}} [ 553.880343] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] Configuration options gathered from: {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2805}} [ 553.880460] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 553.880675] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 553.880798] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ================================================================================ {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 553.880990] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] allow_resize_to_same_host = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881173] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] arq_binding_timeout = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881299] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] backdoor_port = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881419] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] backdoor_socket = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881573] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] block_device_allocate_retries = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881725] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] block_device_allocate_retries_interval = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.881876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cell_worker_thread_pool_size = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882043] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cert = self.pem {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882203] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882362] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute_monitors = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882522] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] config_dir = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882680] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] config_drive_format = iso9660 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882802] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.882953] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] config_source = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883121] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] console_host = devstack {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883287] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] control_exchange = nova {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883440] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cpu_allocation_ratio = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883589] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] daemon = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883742] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] debug = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.883885] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_access_ip_network_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884046] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_availability_zone = nova {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884191] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_ephemeral_format = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884338] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_green_pool_size = 1000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884592] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884748] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_schedule_zone = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.884903] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] default_thread_pool_size = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885060] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] disk_allocation_ratio = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885214] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] enable_new_services = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885362] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] flat_injected = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885530] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] force_config_drive = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885660] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] force_raw_images = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885986] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] graceful_shutdown_timeout = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.885986] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] heal_instance_info_cache_interval = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.886184] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] host = cpu-1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.886356] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.886512] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] initial_disk_allocation_ratio = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.886663] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] initial_ram_allocation_ratio = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.886875] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887041] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_build_timeout = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887196] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_delete_interval = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887374] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_format = [instance: %(uuid)s] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887537] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_name_template = instance-%08x {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887690] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_usage_audit = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887848] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_usage_audit_period = month {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.887999] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888166] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] instances_path = /opt/stack/data/nova/instances {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888319] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] internal_service_availability_zone = internal {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888466] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] key = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888614] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] live_migration_retry_count = 30 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888769] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_color = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.888920] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_config_append = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889085] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889269] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_dir = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889412] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_file = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889532] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_options = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889682] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_rotate_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889839] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_rotate_interval_type = days {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.889994] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] log_rotation_type = none {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890130] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890255] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890407] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890555] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890674] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890820] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] long_rpc_timeout = 1800 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.890968] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_concurrent_builds = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891129] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_concurrent_live_migrations = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891278] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_concurrent_snapshots = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891425] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_local_block_devices = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891571] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_logfile_count = 30 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891715] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] max_logfile_size_mb = 200 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.891861] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] maximum_instance_delete_attempts = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892014] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] migrate_max_retries = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892178] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] mkisofs_cmd = genisoimage {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892367] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] my_block_storage_ip = 10.180.1.21 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892492] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] my_ip = 10.180.1.21 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892681] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] my_shared_fs_storage_ip = 10.180.1.21 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.892836] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] network_allocate_retries = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893024] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893178] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] osapi_compute_unique_server_name_scope = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893329] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] password_length = 12 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893481] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] periodic_enable = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893627] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] periodic_fuzzy_delay = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893779] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] pointer_model = usbtablet {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.893931] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] preallocate_images = none {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894092] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] publish_errors = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894213] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] pybasedir = /opt/stack/nova {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894360] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ram_allocation_ratio = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894507] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rate_limit_burst = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894660] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rate_limit_except_level = CRITICAL {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894806] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rate_limit_interval = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.894953] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reboot_timeout = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895114] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reclaim_instance_interval = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895267] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] record = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895417] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reimage_timeout_per_gb = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895602] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] report_interval = 120 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895759] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rescue_timeout = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.895906] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reserved_host_cpus = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896064] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reserved_host_disk_mb = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896212] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reserved_host_memory_mb = 512 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896360] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] reserved_huge_pages = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896509] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] resize_confirm_window = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896731] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] resize_fs_using_block_device = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.896890] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] resume_guests_state_on_host_boot = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897067] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897228] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] rpc_response_timeout = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897397] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] run_external_periodic_tasks = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897591] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] running_deleted_instance_action = reap {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897702] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] running_deleted_instance_poll_interval = 1800 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897850] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] running_deleted_instance_timeout = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.897995] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler_instance_sync_interval = 120 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898165] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_down_time = 720 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898322] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] servicegroup_driver = db {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898471] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] shell_completion = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898619] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] shelved_offload_time = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898766] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] shelved_poll_interval = 3600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.898919] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] shutdown_timeout = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899084] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] source_is_ipv6 = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899257] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ssl_only = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899504] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899666] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] sync_power_state_interval = 600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899819] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] sync_power_state_pool_size = 1000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.899972] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] syslog_log_facility = LOG_USER {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900134] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] tempdir = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900305] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] thread_pool_statistic_period = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900464] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] timeout_nbd = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900618] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] transport_url = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900765] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] update_resources_interval = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.900910] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_cow_images = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901076] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_journal = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901228] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_json = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901376] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_rootwrap_daemon = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901520] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_stderr = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901684] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] use_syslog = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901831] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vcpu_pin_set = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.901984] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plugging_is_fatal = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902149] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plugging_timeout = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902301] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] virt_mkfs = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902450] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] volume_usage_poll_interval = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902599] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] watch_log_file = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902752] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] web = /usr/share/spice-html5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2817}} [ 553.902927] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_concurrency.disable_process_locking = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.903520] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.903711] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_brick.lock_path = /opt/stack/data/os_brick {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.903874] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904064] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904239] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904396] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904555] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904713] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.904863] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905030] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.compute_link_prefix = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905194] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905350] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.dhcp_domain = novalocal {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905502] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.enable_instance_password = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905648] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.glance_link_prefix = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905796] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.905951] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906113] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.instance_list_per_project_cells = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906262] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.list_records_by_skipping_down_cells = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906410] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.local_metadata_per_cell = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906557] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.max_limit = 1000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906705] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.metadata_cache_expiration = 15 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.906861] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.neutron_default_project_id = default {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907023] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.response_validation = warn {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907177] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.use_neutron_default_nets = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907355] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907514] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907698] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907816] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.907967] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_dynamic_targets = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908131] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_jsonfile_path = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908295] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908476] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.backend = dogpile.cache.memcached {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908631] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.backend_argument = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908775] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.backend_expiration_time = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.908929] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.config_prefix = cache.oslo {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909089] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.debug_cache_backend = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909248] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.enable_retry_client = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909415] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.enable_socket_keepalive = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909573] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.enabled = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909721] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.enforce_fips_mode = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.909870] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.expiration_time = 600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910040] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.hashclient_dead_timeout = 60.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910195] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.hashclient_retry_attempts = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910374] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.hashclient_retry_timeout = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910528] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_dead_retry = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910674] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_password = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910825] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.910982] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911146] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_pool_maxsize = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911298] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911449] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_sasl_enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911614] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911766] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_socket_timeout = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.911910] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.memcache_username = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912079] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.proxies = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912232] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_db = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912378] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_password = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912532] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_sentinel_service_name = mymaster {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912691] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912844] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_server = localhost:6379 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.912993] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_socket_timeout = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913157] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.redis_username = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913305] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.retry_attempts = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913457] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.retry_delay = 0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913607] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.socket_keepalive_count = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913753] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.socket_keepalive_idle = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.913897] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.socket_keepalive_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914051] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.tls_allowed_ciphers = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914199] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.tls_cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914340] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.tls_certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914487] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.tls_enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914630] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cache.tls_keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914781] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.914935] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.auth_type = password {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915093] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915254] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.catalog_info = volumev3::publicURL {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915400] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915548] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915691] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.cross_az_attach = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915838] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.debug = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.915984] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.endpoint_template = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916146] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.http_retries = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916293] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916438] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916591] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.os_region_name = RegionOne {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916737] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.916879] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cinder.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917046] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917195] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.cpu_dedicated_set = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917339] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.cpu_shared_set = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917492] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.image_type_exclude_list = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917636] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917797] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.max_concurrent_disk_ops = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.917924] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.max_disk_devices_to_attach = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918080] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918237] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918390] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.resource_provider_association_refresh = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918538] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918682] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.shutdown_retry_interval = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.918847] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919012] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] conductor.workers = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919180] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] console.allowed_origins = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919353] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] console.ssl_ciphers = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919517] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] console.ssl_minimum_version = default {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919672] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] consoleauth.enforce_session_timeout = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919827] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] consoleauth.token_ttl = 600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.919983] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920142] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920348] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920474] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920620] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920761] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.920906] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921059] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921206] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921347] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921488] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921629] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921768] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.921918] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.service_type = accelerator {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922072] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922217] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922357] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922499] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922660] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922804] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] cyborg.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.922957] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.asyncio_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923115] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.asyncio_slave_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923270] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.backend = sqlalchemy {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923426] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923574] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.connection_debug = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923725] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.connection_parameters = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.923876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.connection_recycle_time = 3600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924032] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.connection_trace = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924181] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.db_inc_retry_interval = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924330] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.db_max_retries = 20 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924479] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.db_max_retry_interval = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924624] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.db_retry_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924771] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.max_overflow = 50 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.924917] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.max_pool_size = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925073] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.max_retries = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925230] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925377] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.mysql_wsrep_sync_wait = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925518] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.pool_timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925663] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.retry_interval = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925805] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.slave_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.925949] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.sqlite_synchronous = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926108] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] database.use_db_reconnect = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926260] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.asyncio_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926408] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.asyncio_slave_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926561] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.backend = sqlalchemy {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926712] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.926861] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.connection_debug = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927051] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.connection_parameters = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927197] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.connection_recycle_time = 3600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927325] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.connection_trace = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927474] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.db_inc_retry_interval = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927622] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.db_max_retries = 20 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927772] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.db_max_retry_interval = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.927921] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.db_retry_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928080] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.max_overflow = 50 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928230] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.max_pool_size = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928378] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.max_retries = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928530] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928675] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928819] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.pool_timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.928966] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.retry_interval = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929121] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.slave_connection = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929291] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] api_database.sqlite_synchronous = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929465] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] devices.enabled_mdev_types = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929632] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929790] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ephemeral_storage_encryption.default_format = luks {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.929939] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ephemeral_storage_encryption.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930102] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930288] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.api_servers = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930458] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930610] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930763] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.930910] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931065] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931219] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.debug = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931374] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.default_trusted_certificate_ids = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931527] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.enable_certificate_validation = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931677] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.enable_rbd_download = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931825] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.931978] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932143] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932294] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932443] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932592] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.num_retries = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932747] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.rbd_ceph_conf = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.932896] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.rbd_connect_timeout = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933063] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.rbd_pool = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933223] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.rbd_user = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933372] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933517] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933661] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933814] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.service_type = image {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.933962] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934120] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934274] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934422] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934587] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934738] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.verify_glance_signatures = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.934885] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] glance.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935052] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] guestfs.debug = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935213] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935362] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.auth_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935510] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935655] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935805] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.935949] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936105] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936254] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936407] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936549] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936694] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936838] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.936983] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937144] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937290] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937444] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.service_type = shared-file-system {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937595] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.share_apply_policy_timeout = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937742] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.937888] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.938048] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.938199] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.938364] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.938514] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] manila.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.938668] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] mks.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939007] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939192] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.manager_interval = 2400 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939375] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.precache_concurrency = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939538] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.remove_unused_base_images = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939697] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.939851] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940021] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] image_cache.subdirectory_name = _base {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940189] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.api_max_retries = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940373] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.api_retry_interval = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940527] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940680] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.auth_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940827] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.940972] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941137] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941287] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.conductor_group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941434] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941579] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941722] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.941869] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942027] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942180] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942326] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942480] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.peer_list = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942626] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942773] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.942919] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.serial_console_state_timeout = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943154] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943327] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.service_type = baremetal {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943478] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.shard = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943628] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943777] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.943922] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944076] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944247] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944396] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ironic.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944563] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944721] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] key_manager.fixed_key = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.944887] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945046] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.barbican_api_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945195] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.barbican_endpoint = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945351] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.barbican_endpoint_type = public {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945497] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.barbican_region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945640] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945783] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.945929] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946085] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946231] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946381] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.number_of_retries = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946526] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.retry_delay = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946670] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.send_service_user_token = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.946814] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.947012] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.947269] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.verify_ssl = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.947500] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican.verify_ssl_path = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.947748] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.948010] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.auth_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.948282] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.948542] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.948793] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949065] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949311] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949497] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949650] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] barbican_service_user.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949807] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.approle_role_id = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.949957] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.approle_secret_id = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950130] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.kv_mountpoint = secret {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950280] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.kv_path = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950433] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.kv_version = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950580] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.namespace = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950729] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.root_token_id = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.950876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.ssl_ca_crt_file = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951040] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.timeout = 60.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951194] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.use_ssl = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951349] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951506] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951653] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951804] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.951948] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952106] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952254] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952404] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952549] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952692] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952835] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.952979] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953139] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953287] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953446] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.service_type = identity {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953593] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953738] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.953887] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.954047] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.954217] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.954365] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] keystone.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.954541] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.ceph_mount_options = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.954849] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.ceph_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955030] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.connection_uri = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955188] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_mode = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955343] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_model_extra_flags = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955503] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_models = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955661] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_power_governor_high = performance {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955816] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_power_governor_low = powersave {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.955966] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_power_management = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956137] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956291] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.device_detach_attempts = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956442] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.device_detach_timeout = 20 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956591] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.disk_cachemodes = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956739] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.disk_prefix = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.956889] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.enabled_perf_events = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957055] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.file_backed_memory = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957215] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.gid_maps = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957362] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.hw_disk_discard = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957510] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.hw_machine_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957665] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_rbd_ceph_conf = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957815] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.957963] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958136] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_rbd_glance_store_name = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958289] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_rbd_pool = rbd {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958444] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_type = default {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958589] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.images_volume_group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958738] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.inject_key = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.958886] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.inject_partition = -2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959042] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.inject_password = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959195] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.iscsi_iface = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959372] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.iser_use_multipath = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959528] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_bandwidth = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959677] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959824] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_downtime = 500 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.959971] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960130] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960285] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_inbound_addr = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960436] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_parallel_connections = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960582] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960729] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_permit_post_copy = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.960873] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_scheme = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961040] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_timeout_action = abort {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961198] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_tunnelled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961344] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_uri = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961491] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.live_migration_with_native_tls = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961639] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.max_queues = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961786] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.961991] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.962153] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.nfs_mount_options = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.962474] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.962718] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.962895] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_iser_scan_tries = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.963060] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_memory_encrypted_guests = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.963411] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.963585] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_pcie_ports = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.963746] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.num_volume_scan_tries = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.963904] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.pmem_namespaces = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.964066] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.quobyte_client_cfg = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.964364] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.964539] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rbd_connect_timeout = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.964693] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.964857] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965044] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rbd_secret_uuid = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965203] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rbd_user = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965355] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965516] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.remote_filesystem_transport = ssh {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965701] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rescue_image_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.965933] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rescue_kernel_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.966120] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rescue_ramdisk_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.966284] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.966435] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.rx_queue_size = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.966592] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.smbfs_mount_options = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.966878] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967055] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.snapshot_compression = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967209] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.snapshot_image_format = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967444] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967601] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.sparse_logical_volumes = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967765] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.supported_tpm_secret_security = ['user', 'host'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.967915] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.swtpm_enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968081] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.swtpm_group = tss {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968238] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.swtpm_user = tss {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968395] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.sysinfo_serial = unique {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968537] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.tb_cache_size = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968681] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.tx_queue_size = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968830] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.uid_maps = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.968977] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.use_default_aio_mode_for_volumes = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969139] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.use_virtio_for_bridges = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969322] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.virt_type = kvm {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969490] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.volume_clear = zero {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969640] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.volume_clear_size = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969789] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.volume_enforce_multipath = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.969938] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.volume_use_multipath = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.970098] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_cache_path = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.970270] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.970435] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_mount_group = qemu {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.970588] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_mount_opts = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.970742] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971010] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971187] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.vzstorage_mount_user = stack {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971341] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971503] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971655] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.auth_type = password {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.971799] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972011] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972194] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972343] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972489] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972645] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.default_floating_pool = public {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972795] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.972947] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.extension_sync_interval = 600 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973108] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.http_retries = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973258] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973406] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973552] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973706] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.973852] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974015] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.ovs_bridge = br-int {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974175] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.physnets = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974332] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.region_name = RegionOne {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974480] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974632] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.service_metadata_proxy = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974778] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.974932] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.service_type = network {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975091] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975240] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975386] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975532] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975705] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.975854] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] neutron.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976024] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.bdms_in_notifications = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976190] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.default_level = INFO {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976342] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.include_share_mapping = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976503] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.notification_format = unversioned {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976651] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.notify_on_state_change = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976814] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.976972] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] pci.alias = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977143] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] pci.device_spec = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977345] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] pci.report_in_placement = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977532] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977697] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.auth_type = password {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977849] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.977995] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978156] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978315] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978453] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978595] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978738] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.default_domain_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.978879] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.default_domain_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979029] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.domain_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979177] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.domain_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979347] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979509] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979655] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979799] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.979943] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980109] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.password = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980259] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.project_domain_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980415] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.project_domain_name = Default {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980565] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.project_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980722] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.project_name = service {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.980876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.region_name = RegionOne {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981033] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981185] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981338] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.service_type = placement {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981486] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981630] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981773] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.981917] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.system_scope = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982069] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982214] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.trust_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982358] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.user_domain_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982509] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.user_domain_name = Default {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982650] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.user_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982805] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.username = nova {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.982965] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983122] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] placement.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983296] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.cores = 20 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983447] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.count_usage_from_placement = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983602] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983754] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.injected_file_content_bytes = 10240 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.983906] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.injected_file_path_length = 255 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984065] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.injected_files = 5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984218] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.instances = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984397] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.key_pairs = 100 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984556] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.metadata_items = 128 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984706] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.ram = 51200 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.984853] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.recheck_quota = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985018] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.server_group_members = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985167] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.server_groups = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985357] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.unified_limits_resource_list = ['servers'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985519] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] quota.unified_limits_resource_strategy = require {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985676] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985824] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.985970] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.image_metadata_prefilter = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986130] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986282] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.max_attempts = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986431] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.max_placement_results = 1000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986579] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986722] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.query_placement_for_image_type_support = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.986867] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987043] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] scheduler.workers = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987207] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987364] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987530] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987683] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987831] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.987980] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988143] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988322] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988486] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.host_subset_size = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988629] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988774] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.988924] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.image_props_weight_multiplier = 0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989094] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.image_props_weight_setting = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989284] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989417] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.isolated_hosts = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989568] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.isolated_images = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989730] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.989876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990035] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990188] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.pci_in_placement = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990339] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990489] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990637] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990782] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.990928] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991087] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991234] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.track_instance_changes = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991395] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991552] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] metrics.required = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991700] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] metrics.weight_multiplier = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991846] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.991993] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] metrics.weight_setting = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.992307] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.992470] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.992634] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.port_range = 10000:20000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.992790] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.992944] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993117] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] serial_console.serialproxy_port = 6083 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993273] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993432] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.auth_type = password {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993577] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993721] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.993865] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994012] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994158] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994313] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.send_service_user_token = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994461] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994603] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] service_user.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994755] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.agent_enabled = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.994900] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.995208] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.995403] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.995562] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.html5proxy_port = 6082 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.995712] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.image_compression = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.995859] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.jpeg_compression = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996008] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.playback_compression = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996167] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.require_secure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996320] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.server_listen = 127.0.0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996508] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996778] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.spice_direct_proxy_base_url = http://127.0.0.1:13002/nova {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.996940] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.streaming_mode = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997103] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] spice.zlib_compression = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997263] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] upgrade_levels.baseapi = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997423] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] upgrade_levels.compute = auto {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997624] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] upgrade_levels.conductor = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997774] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] upgrade_levels.scheduler = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.997929] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998091] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.auth_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998241] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998387] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998537] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998683] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998823] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.998970] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999126] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vendordata_dynamic_auth.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999293] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.api_retry_count = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999437] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.ca_file = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999592] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.cache_prefix = devstack-image-cache {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999743] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.cluster_name = testcl1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 553.999891] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.connection_pool_size = 10 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000051] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.console_delay_seconds = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000210] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.datastore_regex = ^datastore.* {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000413] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000586] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.host_password = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000741] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.host_port = 443 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.000896] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.host_username = administrator@vsphere.local {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001064] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.insecure = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001216] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.integration_bridge = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001369] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.maximum_objects = 100 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001518] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.pbm_default_policy = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001670] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.pbm_enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001813] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.pbm_wsdl_location = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.001967] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002125] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.serial_port_proxy_uri = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002270] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.serial_port_service_uri = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002428] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.task_poll_interval = 0.5 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002585] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.use_linked_clone = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002736] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.vnc_keymap = en-us {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.002884] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.vnc_port = 5900 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003052] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vmware.vnc_port_total = 10000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003227] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.auth_schemes = ['none'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003388] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003654] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003822] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.003979] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.novncproxy_port = 6080 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004178] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.server_listen = 127.0.0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004346] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004497] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.vencrypt_ca_certs = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004643] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.vencrypt_client_cert = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004787] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vnc.vencrypt_client_key = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.004947] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005119] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_deep_image_inspection = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005273] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005422] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005567] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005712] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.disable_rootwrap = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005856] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.enable_numa_live_migration = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.005999] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006159] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006306] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006451] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.libvirt_disable_apic = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006596] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006742] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.006887] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007041] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007190] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007340] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007487] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007632] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007778] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.007928] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008106] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008258] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] wsgi.secure_proxy_ssl_header = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008412] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] zvm.ca_file = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008555] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] zvm.cloud_connector_url = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008836] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.008995] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] zvm.reachable_timeout = 300 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009165] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009328] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009491] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.connection_string = messaging:// {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009642] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.enabled = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009795] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.es_doc_type = notification {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.009942] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.es_scroll_size = 10000 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010112] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.es_scroll_time = 2m {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010281] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.filter_error_trace = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010430] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.hmac_keys = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010583] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.sentinel_service_name = mymaster {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010735] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.socket_timeout = 0.1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.010880] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.trace_requests = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011035] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler.trace_sqlalchemy = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011209] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler_jaeger.process_tags = {} {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011353] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler_jaeger.service_name_prefix = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011501] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] profiler_otlp.service_name_prefix = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011658] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011802] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.011946] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012105] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012251] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012397] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012540] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012684] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012828] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.012980] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013136] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013289] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013441] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013590] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013735] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.kombu_reconnect_splay = 0.0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.013884] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014042] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014191] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014342] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014491] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014635] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014780] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.014922] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015079] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015234] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015379] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015523] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015669] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015812] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.015957] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016114] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016269] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016423] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016569] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016720] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.016870] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.ssl_version = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017027] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017201] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017353] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_notifications.retry = -1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017512] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017665] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_messaging_notifications.transport_url = **** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017818] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.auth_section = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.017966] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.auth_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018123] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.cafile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018267] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.certfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018415] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.collect_timing = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018564] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.connect_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018708] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.connect_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.018873] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_id = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019010] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_interface = publicURL {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019160] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_override = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019316] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019449] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019589] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.endpoint_service_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019732] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.insecure = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.019871] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.keyfile = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020016] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.max_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020159] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.min_version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020302] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.region_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020446] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.retriable_status_codes = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020586] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.service_name = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020725] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.service_type = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.020868] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.split_loggers = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021015] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.status_code_retries = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021163] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.status_code_retry_delay = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021303] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.timeout = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021445] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.valid_interfaces = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021584] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_limit.version = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021731] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_reports.file_event_handler = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.021876] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022027] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] oslo_reports.log_dir = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022184] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022326] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022468] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022612] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.log_daemon_traceback = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022759] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.022907] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023063] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023220] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023367] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023511] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023658] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.log_daemon_traceback = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023807] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.023953] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024110] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] vif_plug_ovs_privileged.user = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024266] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.flat_interface = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024430] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024587] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024741] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.024895] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025062] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025216] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025365] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025529] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025684] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.isolate_vif = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025836] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.025987] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026154] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026309] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.ovsdb_interface = native {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026459] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] os_vif_ovs.per_port_bridge = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026611] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.capabilities = [21, 2] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026755] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.026894] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.helper_command = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027052] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.log_daemon_traceback = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027205] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027354] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027498] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] privsep_osbrick.user = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027654] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027796] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.group = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.027939] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.helper_command = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.028095] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.log_daemon_traceback = False {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.028247] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.028396] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.028537] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] nova_sys_admin.user = None {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2824}} [ 554.028656] env[66641]: DEBUG oslo_service.backend._eventlet.service [None req-cb64843d-8977-4446-8ba2-377ff919f372 None None] ******************************************************************************** {{(pid=66641) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2828}} [ 554.029103] env[66641]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 554.533324] env[66641]: WARNING nova.virt.vmwareapi.driver [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 554.534091] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Getting list of instances from cluster (obj){ [ 554.534091] env[66641]: value = "domain-c8" [ 554.534091] env[66641]: _type = "ClusterComputeResource" [ 554.534091] env[66641]: } {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 554.535247] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b88aecf1-484f-4471-96a1-e90b6a2ddca6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.546013] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Got total of 0 instances {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 554.546638] env[66641]: INFO nova.virt.node [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Generated node identity 750ffd2d-5e46-4240-a614-995f2be7c9cb [ 554.546946] env[66641]: INFO nova.virt.node [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Wrote node identity 750ffd2d-5e46-4240-a614-995f2be7c9cb to /opt/stack/data/n-cpu-1/compute_id [ 555.050580] env[66641]: WARNING nova.compute.manager [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Compute nodes ['750ffd2d-5e46-4240-a614-995f2be7c9cb'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 556.057072] env[66641]: INFO nova.compute.manager [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 557.063608] env[66641]: WARNING nova.compute.manager [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 557.064042] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 557.064042] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 557.064328] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 557.064328] env[66641]: DEBUG nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 557.065210] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1aa1f58-930e-45c5-a1dd-c97e827085dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.073704] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55709f86-5737-4cb0-be38-d2c2b3327e31 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.087633] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c5504ce-c9d2-4b44-acf9-ed4e4622647f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.094900] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe4f480-f592-477a-8268-59c107306db5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.123364] env[66641]: DEBUG nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180554MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 557.123549] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 557.123715] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 557.626078] env[66641]: WARNING nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] No compute node record for cpu-1:750ffd2d-5e46-4240-a614-995f2be7c9cb: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 750ffd2d-5e46-4240-a614-995f2be7c9cb could not be found. [ 558.130413] env[66641]: INFO nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 750ffd2d-5e46-4240-a614-995f2be7c9cb [ 559.638608] env[66641]: DEBUG nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 559.639059] env[66641]: DEBUG nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '0'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 559.799290] env[66641]: INFO nova.scheduler.client.report [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] [req-3c31d9ff-e1a5-4330-9b9d-da93f60bc031] Created resource provider record via placement API for resource provider with UUID 750ffd2d-5e46-4240-a614-995f2be7c9cb and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 559.817090] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d094c6-3584-444f-92dc-2e504173459d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.825231] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b47932-90be-49f9-a7cd-bbf33eee855a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.856343] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51791e98-7b45-4386-9437-4524073e634c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.863764] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e193d819-20da-4ce7-8284-4c16020f8e0f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.878095] env[66641]: DEBUG nova.compute.provider_tree [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 560.414618] env[66641]: DEBUG nova.scheduler.client.report [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Updated inventory for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 560.414852] env[66641]: DEBUG nova.compute.provider_tree [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Updating resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb generation from 0 to 1 during operation: update_inventory {{(pid=66641) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 560.414986] env[66641]: DEBUG nova.compute.provider_tree [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 560.465566] env[66641]: DEBUG nova.compute.provider_tree [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Updating resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb generation from 1 to 2 during operation: update_traits {{(pid=66641) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 560.970355] env[66641]: DEBUG nova.compute.resource_tracker [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 560.970819] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.847s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 560.970819] env[66641]: DEBUG nova.service [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Creating RPC server for service compute {{(pid=66641) start /opt/stack/nova/nova/service.py:177}} [ 560.986097] env[66641]: DEBUG nova.service [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] Join ServiceGroup membership for this service compute {{(pid=66641) start /opt/stack/nova/nova/service.py:194}} [ 560.986312] env[66641]: DEBUG nova.servicegroup.drivers.db [None req-6719d073-d0a7-4fea-850f-3905b955e570 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=66641) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 579.988695] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._sync_power_states {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 580.491689] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Getting list of instances from cluster (obj){ [ 580.491689] env[66641]: value = "domain-c8" [ 580.491689] env[66641]: _type = "ClusterComputeResource" [ 580.491689] env[66641]: } {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 580.492932] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba6f450-9a9e-4664-a3eb-c8f18661682a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.501994] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Got total of 0 instances {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 580.502219] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 580.502512] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Getting list of instances from cluster (obj){ [ 580.502512] env[66641]: value = "domain-c8" [ 580.502512] env[66641]: _type = "ClusterComputeResource" [ 580.502512] env[66641]: } {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 580.503450] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ad684be-bc75-4384-93cc-c4c96325946b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.510841] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Got total of 0 instances {{(pid=66641) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 607.328471] env[66641]: INFO nova.utils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] The default thread pool MainProcess.default is initialized [ 607.331181] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.331401] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 607.530313] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.530641] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 607.553031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 607.553031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 607.834593] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 608.034579] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 608.057622] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 608.382510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.382824] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 608.385077] env[66641]: INFO nova.compute.claims [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.579865] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 608.592797] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.192516] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.193087] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.193312] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.193491] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.193667] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.193855] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.194055] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.194226] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 609.194379] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 609.568685] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b92f7fc-3531-4253-aad0-e427aafbf81e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.581855] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55398d67-5cd8-4e94-8490-41706b7e7a09 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.615733] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e2e50b-332d-41c7-b6bc-0e9b9d9602b6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.624464] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ab62ff7-a9b5-48eb-bb44-b89363efb24c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.643669] env[66641]: DEBUG nova.compute.provider_tree [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.702354] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.781502] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 609.781884] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.145466] env[66641]: DEBUG nova.scheduler.client.report [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 610.287875] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 610.650888] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 610.651537] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 610.655752] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.078s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 610.657130] env[66641]: INFO nova.compute.claims [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 610.815837] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.165851] env[66641]: DEBUG nova.compute.utils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 611.167980] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 611.168445] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 611.169307] env[66641]: WARNING neutronclient.v2_0.client [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.170969] env[66641]: WARNING neutronclient.v2_0.client [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 611.175020] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 611.175020] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 611.654576] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 611.654938] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 611.681877] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 611.837574] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f679a2c6-5745-40de-951d-08abd363b893 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.848301] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06054804-98f3-46af-95bc-10a6a9dd8e22 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.883968] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189df14c-3670-486f-9c89-e12789ea3aef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.893395] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ad69fb-3bda-4163-9e1d-1652a72e4713 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.909611] env[66641]: DEBUG nova.compute.provider_tree [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.161582] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 612.413452] env[66641]: DEBUG nova.scheduler.client.report [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 612.655779] env[66641]: DEBUG nova.policy [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd3266c772594eeb8b28bbb070f1c70a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37c028dc169a444fa76652ff45efc917', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 612.695825] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 612.696847] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 612.745861] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 612.746135] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 612.746290] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 612.746463] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 612.747417] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 612.747417] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 612.747417] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 612.747417] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 612.747737] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 612.747737] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 612.747804] env[66641]: DEBUG nova.virt.hardware [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 612.748756] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7cd6a1-a9ce-42e6-86a3-6eed55e00488 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.760645] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926eed03-cbf1-4579-b970-38a465c497e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.782508] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9457a57-5dff-48d1-b9e8-92fcb1866d03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.918978] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 612.919484] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 612.922217] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.330s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 612.923630] env[66641]: INFO nova.compute.claims [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 613.428237] env[66641]: DEBUG nova.compute.utils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 613.432968] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 613.765769] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 613.768109] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 613.913591] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Successfully created port: 0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 613.933268] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 614.127797] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d9ecda-91c7-4ea4-9da7-df0aa47a0433 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.136604] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74206d5e-4929-446d-83d8-80c0370f63d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.168541] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e11aedb8-3c2b-4d1e-a1cf-9913aea2dfaf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.178600] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb59d9fb-2ab9-450f-aede-3ddda33eb0d6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.195602] env[66641]: DEBUG nova.compute.provider_tree [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.699525] env[66641]: DEBUG nova.scheduler.client.report [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 614.949270] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 614.980109] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 614.980109] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 614.980109] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 614.980343] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 614.980591] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 614.980945] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 614.981268] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 614.981540] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 614.981834] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 614.982183] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 614.984018] env[66641]: DEBUG nova.virt.hardware [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 614.984018] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8660860-3805-4dc7-a395-ddcf8477397b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.994382] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9886a3-db8b-42e2-8398-32fbc2ca8424 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.012841] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 615.022672] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 615.022672] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f80a5ec9-6db1-4a35-afeb-3fd5733d1cf2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.037909] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Created folder: OpenStack in parent group-v4. [ 615.038458] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating folder: Project (c00dc422cedf4cec85b175435c6d19fc). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 615.038762] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c499dfe3-26a0-4eb2-b7c3-35defb79c9c8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.050584] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Created folder: Project (c00dc422cedf4cec85b175435c6d19fc) in parent group-v1000566. [ 615.050791] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating folder: Instances. Parent ref: group-v1000567. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 615.051144] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e045bf6-59aa-4837-87e6-aaa4acf7814b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.062193] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Created folder: Instances in parent group-v1000567. [ 615.062471] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 615.062668] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 615.062872] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c7ae9ea9-2439-4078-8b04-f470639047da {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.082078] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 615.082078] env[66641]: value = "task-5145530" [ 615.082078] env[66641]: _type = "Task" [ 615.082078] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.092318] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145530, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.206644] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.284s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.206768] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 615.210792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.507s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.210792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 615.210792] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 615.210792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.394s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 615.211348] env[66641]: INFO nova.compute.claims [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 615.216124] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba663df2-fc8d-4638-8d97-f316ee5e47eb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.229755] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08c3263c-d2bf-449f-9e26-9dfa2bd61683 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.247929] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57728ab-6cfd-4b97-836b-d3a98193f7a1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.257695] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fc864a-9d1e-451a-920b-bb3db7ef3ecd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.292211] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180563MB free_disk=96GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 615.292344] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 615.594435] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145530, 'name': CreateVM_Task, 'duration_secs': 0.359065} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 615.594435] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 615.597055] env[66641]: DEBUG oslo_vmware.service [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9bccaca-4c33-400b-987d-0ddb24d9853b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.606073] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.606073] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 615.606557] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 615.606814] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3042ae15-e7a3-4255-9eeb-f33f5c2c4ed7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.612304] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 615.612304] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5291e284-4bf3-b887-7263-225de454c8d9" [ 615.612304] env[66641]: _type = "Task" [ 615.612304] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 615.620919] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5291e284-4bf3-b887-7263-225de454c8d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 615.717243] env[66641]: DEBUG nova.compute.utils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 615.719628] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 615.719628] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 615.719628] env[66641]: WARNING neutronclient.v2_0.client [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.719628] env[66641]: WARNING neutronclient.v2_0.client [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 615.720143] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 615.720529] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 615.821494] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Successfully updated port: 0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 616.130022] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 616.130022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 616.130022] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.130022] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 616.130441] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 616.130441] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19829386-e271-4f76-99e7-0a495ee36db1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.149721] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 616.150317] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 616.151201] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20a6f10-7263-4ed5-b979-9d5de53b5338 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.162257] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfd3fb6c-4daa-4c80-988e-ad504977798c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.167409] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 616.167409] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5206cbde-4ef5-630b-a6ad-6bda04660744" [ 616.167409] env[66641]: _type = "Task" [ 616.167409] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 616.179327] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5206cbde-4ef5-630b-a6ad-6bda04660744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 616.228095] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 616.326308] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.326308] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquired lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 616.326308] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 616.371444] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371b7041-3549-4a50-a554-e64339336759 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.383937] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9afbc8d9-25b0-4799-98e5-126f4b019e39 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.420432] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c148fa33-9b52-42ef-a880-16e49cc196c8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.427012] env[66641]: DEBUG nova.policy [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b298c72117a4e718fb96c91ff24a310', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '09cb6ffdd65e4f579dabc938cadc366d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 616.440125] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 616.440125] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 616.447626] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-598b4a63-8e80-4cd7-8f0e-7663ac4da1a4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.465987] env[66641]: DEBUG nova.compute.provider_tree [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.680135] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Preparing fetch location {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 616.681063] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating directory with path [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 616.681063] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-46dcc42b-a019-434b-b543-ce230beefcef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.702777] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Created directory with path [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 616.702777] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Fetch image to [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 616.702946] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Downloading image file data cd910b37-6707-4868-b172-79fffc590a51 to [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk on the data store datastore2 {{(pid=66641) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 616.703761] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81927cec-5953-4142-8607-555c33110848 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.712868] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d39e262f-5eb1-46d2-ab07-a3dc4449f41a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.723549] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27ae541-b531-4353-b1fb-7aca33160a77 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.759856] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16370ee6-8d3e-4bc2-8e07-0fb2de8317a3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.767144] env[66641]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-f8de5b8d-0bd6-4967-84cd-dae8d7dced95 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.802575] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Downloading image file data cd910b37-6707-4868-b172-79fffc590a51 to the data store datastore2 {{(pid=66641) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 616.829625] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 616.829946] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 616.835510] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 616.897304] env[66641]: DEBUG oslo_vmware.rw_handles [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=66641) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 616.961023] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 616.968868] env[66641]: DEBUG nova.scheduler.client.report [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 617.230616] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Successfully created port: 361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 617.263880] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 617.308395] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 617.308679] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 617.308854] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 617.309724] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 617.309843] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 617.310086] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 617.310319] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.310473] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 617.310630] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 617.310900] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 617.310971] env[66641]: DEBUG nova.virt.hardware [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 617.312138] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91685eea-8b31-4782-bdbd-21b900d4763d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.326856] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a90564-0d51-4072-92a8-eaba6156c016 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.477428] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.267s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 617.477974] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 617.481963] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.787s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 617.486434] env[66641]: INFO nova.compute.claims [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 617.513974] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 617.616278] env[66641]: DEBUG oslo_vmware.rw_handles [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Completed reading data from the image iterator. {{(pid=66641) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 617.616278] env[66641]: DEBUG oslo_vmware.rw_handles [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 617.758523] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Downloaded image file data cd910b37-6707-4868-b172-79fffc590a51 to vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk on the data store datastore2 {{(pid=66641) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 617.759992] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Caching image {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 617.760257] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Copying Virtual Disk [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk to [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 617.760548] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52b8c6b5-faaf-4105-a0a1-b0e1c7064bd9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.771954] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 617.771954] env[66641]: value = "task-5145531" [ 617.771954] env[66641]: _type = "Task" [ 617.771954] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 617.782113] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145531, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 617.988941] env[66641]: DEBUG nova.compute.utils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 617.993369] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 617.993599] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 617.993885] env[66641]: WARNING neutronclient.v2_0.client [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.994667] env[66641]: WARNING neutronclient.v2_0.client [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 617.995621] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 617.995998] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 618.295974] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145531, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.492156] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 618.662952] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13120d9c-af54-4a2e-ac97-477723f89bf0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.673868] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-297e8ad7-4872-44aa-9e2b-714450266996 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.713558] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15db1331-8a0c-42d1-b73c-21115b015284 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.722400] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998aadc5-da07-4f78-afa6-2e6ebd53afff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.737931] env[66641]: DEBUG nova.compute.provider_tree [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 618.782471] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145531, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.731306} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 618.782738] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Copied Virtual Disk [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk to [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 618.782934] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleting the datastore file [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 618.783270] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-999a3e5c-b689-4bbe-b1fe-5cae09e124d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.791304] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 618.791304] env[66641]: value = "task-5145532" [ 618.791304] env[66641]: _type = "Task" [ 618.791304] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 618.807555] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145532, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 618.970920] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 618.971076] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.122185] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Successfully updated port: 361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 619.270758] env[66641]: ERROR nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [req-cce3f74d-86d9-47ec-ad28-7b4cfbd027ae] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 750ffd2d-5e46-4240-a614-995f2be7c9cb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-cce3f74d-86d9-47ec-ad28-7b4cfbd027ae"}]} [ 619.296630] env[66641]: DEBUG nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Refreshing inventories for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 619.304954] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145532, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.027462} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.304954] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 619.305067] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Moving file from [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9/cd910b37-6707-4868-b172-79fffc590a51 to [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51. {{(pid=66641) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 619.305658] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-dbe5029b-f5aa-4e5e-873e-ab4768ca62e5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.314406] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 619.314406] env[66641]: value = "task-5145533" [ 619.314406] env[66641]: _type = "Task" [ 619.314406] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.319371] env[66641]: DEBUG nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating ProviderTree inventory for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 619.319551] env[66641]: DEBUG nova.compute.provider_tree [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 96, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 619.330599] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145533, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.340521] env[66641]: DEBUG nova.policy [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d0eeef5974e470fa2a3b78e7b25bf9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '82383b2f8cf2490ebdf92495d9dc20e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 619.348964] env[66641]: DEBUG nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Refreshing aggregate associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, aggregates: 76783287-4f61-4c75-8858-d1331407c462 {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 619.384615] env[66641]: DEBUG nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Refreshing trait associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, traits: HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 619.404220] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 619.404764] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 619.505584] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 619.555622] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 619.555963] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 619.556118] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 619.556304] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 619.556448] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 619.556646] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 619.556800] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 619.556924] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 619.557091] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 619.557289] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 619.557855] env[66641]: DEBUG nova.virt.hardware [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 619.558381] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5a23de5-eaee-4a2b-a0d1-2008c9d84724 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.564276] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5e1d75-bb72-4668-8ed2-d30babead4b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.574140] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cb197d5-1a25-4b2b-a25d-a75e10056c10 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.593423] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d042043-9ca7-4369-b994-001e4855e3bc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.629301] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.629503] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquired lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 619.629673] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 619.632479] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39230e9b-b038-447b-b9b4-e82520be6389 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.642427] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080da675-9a9e-4048-9bbd-e2db78857391 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.659776] env[66641]: DEBUG nova.compute.provider_tree [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 619.829102] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145533, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.03298} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 619.829527] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] File moved {{(pid=66641) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 619.829744] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Cleaning up location [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 619.829905] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleting the datastore file [datastore2] vmware_temp/9c0ca2a0-6d2e-485f-abe8-bf6b4bae83d9 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 619.830194] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f36fbaeb-0faa-4c2a-8bab-fb6b2226e21b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.838893] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 619.838893] env[66641]: value = "task-5145534" [ 619.838893] env[66641]: _type = "Task" [ 619.838893] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 619.848909] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 619.941878] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Successfully created port: 88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 620.092384] env[66641]: DEBUG nova.network.neutron [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Updating instance_info_cache with network_info: [{"id": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "address": "fa:16:3e:d6:f8:82", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c57e5de-89", "ovs_interfaceid": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 620.134331] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 620.134985] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 620.149070] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 620.225138] env[66641]: DEBUG nova.scheduler.client.report [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updated inventory for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with generation 9 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:975}} [ 620.225664] env[66641]: DEBUG nova.compute.provider_tree [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb generation from 9 to 10 during operation: update_inventory {{(pid=66641) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 620.225923] env[66641]: DEBUG nova.compute.provider_tree [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 620.354194] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.03014} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.354493] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 620.356163] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30ecf1c4-38f6-4ca5-a1da-5e1946aca4c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.363438] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 620.363438] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520687af-29ed-5ad4-43b2-6a97e80aab5a" [ 620.363438] env[66641]: _type = "Task" [ 620.363438] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.377665] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520687af-29ed-5ad4-43b2-6a97e80aab5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.527868] env[66641]: DEBUG nova.compute.manager [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Received event network-vif-plugged-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 620.528067] env[66641]: DEBUG oslo_concurrency.lockutils [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] Acquiring lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.530160] env[66641]: DEBUG oslo_concurrency.lockutils [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.530410] env[66641]: DEBUG oslo_concurrency.lockutils [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 620.530598] env[66641]: DEBUG nova.compute.manager [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] No waiting events found dispatching network-vif-plugged-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 620.530802] env[66641]: WARNING nova.compute.manager [req-7369fa72-ee9f-426d-b9c3-a3a3f29a302b req-e91e2c65-5e47-4ccf-b166-f249cd060ec3 service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Received unexpected event network-vif-plugged-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 for instance with vm_state building and task_state spawning. [ 620.594691] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Releasing lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.595098] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Instance network_info: |[{"id": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "address": "fa:16:3e:d6:f8:82", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c57e5de-89", "ovs_interfaceid": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 620.596026] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d6:f8:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c57e5de-8952-4f18-8e92-c4cc1fb19ef7', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 620.604919] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Creating folder: Project (37c028dc169a444fa76652ff45efc917). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.605266] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c52f8a76-bd28-45fa-a191-b38ec83a3509 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.617942] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Created folder: Project (37c028dc169a444fa76652ff45efc917) in parent group-v1000566. [ 620.618126] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Creating folder: Instances. Parent ref: group-v1000570. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 620.618258] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-404d70df-6f90-47dd-8d9d-e1092fa093b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.629449] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Created folder: Instances in parent group-v1000570. [ 620.629630] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 620.629798] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 620.630165] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6d75c39-6b38-443c-8fd1-775ae9d31c47 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.655085] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 620.655085] env[66641]: value = "task-5145537" [ 620.655085] env[66641]: _type = "Task" [ 620.655085] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.665304] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145537, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 620.736686] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.255s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 620.737763] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 620.740489] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.448s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.874500] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520687af-29ed-5ad4-43b2-6a97e80aab5a, 'name': SearchDatastore_Task, 'duration_secs': 0.012839} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 620.876279] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 620.876279] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 4ebc0f2b-bb79-4941-b8f2-082560c80f8b/4ebc0f2b-bb79-4941-b8f2-082560c80f8b.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 620.878985] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0fdf937b-1ae0-47ac-95ea-d22884ad08b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.881968] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 620.882236] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 620.890473] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 620.890473] env[66641]: value = "task-5145538" [ 620.890473] env[66641]: _type = "Task" [ 620.890473] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 620.900086] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145538, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.020850] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.020850] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.166856] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145537, 'name': CreateVM_Task, 'duration_secs': 0.405655} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.167087] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 621.168090] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.168509] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.173895] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.174726] env[66641]: WARNING openstack [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.244178] env[66641]: DEBUG nova.compute.utils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 621.250452] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 621.250452] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 621.250452] env[66641]: WARNING neutronclient.v2_0.client [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.250624] env[66641]: WARNING neutronclient.v2_0.client [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 621.251417] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.251946] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.259976] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 621.385984] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 621.483982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.484249] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.484586] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 621.488037] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aca80ccf-7f5b-4250-a093-5f5ccdc5ff5a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.496234] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 621.496583] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 621.518157] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 621.518157] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eee185-b113-de3c-49b8-cd3dc2910232" [ 621.518157] env[66641]: _type = "Task" [ 621.518157] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.522728] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145538, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578588} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.526197] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 4ebc0f2b-bb79-4941-b8f2-082560c80f8b/4ebc0f2b-bb79-4941-b8f2-082560c80f8b.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 621.526656] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 621.526882] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-52971ea9-8534-41f5-8e71-5a092b624842 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.535820] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eee185-b113-de3c-49b8-cd3dc2910232, 'name': SearchDatastore_Task, 'duration_secs': 0.009115} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.537126] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 621.538760] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 621.539383] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 621.539383] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.539630] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 621.539722] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 621.540213] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 621.540213] env[66641]: value = "task-5145539" [ 621.540213] env[66641]: _type = "Task" [ 621.540213] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.540441] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-29e974a5-d098-4e39-a9bb-7f1cbc69ab02 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.554550] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145539, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.559883] env[66641]: DEBUG nova.policy [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1eb90a1845bb4360816715d1e0f51b5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '417e1e40aca447229001fa725c82e9ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 621.563098] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 621.563098] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 621.563916] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c26d130-1f38-4ee3-9f48-ed485ee9c092 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.570347] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 621.570347] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5200f348-73d2-3532-2f01-debd62dfc0ab" [ 621.570347] env[66641]: _type = "Task" [ 621.570347] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.585515] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5200f348-73d2-3532-2f01-debd62dfc0ab, 'name': SearchDatastore_Task, 'duration_secs': 0.011139} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 621.585515] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24205a6c-2d72-46fd-815a-12d39e636886 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.592542] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 621.592542] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b51f21-ce05-417a-267c-7d6623f1efaf" [ 621.592542] env[66641]: _type = "Task" [ 621.592542] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 621.604019] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b51f21-ce05-417a-267c-7d6623f1efaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 621.659473] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Successfully updated port: 88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 621.800846] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 25203771-cf4d-4fc1-a1bf-f2a0200296b7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 621.800998] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance ccf46c5c-1c79-4672-ad9f-ea61042097d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 621.801134] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 4ebc0f2b-bb79-4941-b8f2-082560c80f8b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 621.801251] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 621.801360] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 622.055812] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145539, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073769} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.056350] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 622.057133] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf224a5e-3774-45c8-8f76-6fd0afd36c58 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.082959] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 4ebc0f2b-bb79-4941-b8f2-082560c80f8b/4ebc0f2b-bb79-4941-b8f2-082560c80f8b.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 622.083307] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba0ed9c0-0b82-47ee-b2c2-2c9835a964ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.110213] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b51f21-ce05-417a-267c-7d6623f1efaf, 'name': SearchDatastore_Task, 'duration_secs': 0.010425} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.111612] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.111878] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 25203771-cf4d-4fc1-a1bf-f2a0200296b7/25203771-cf4d-4fc1-a1bf-f2a0200296b7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 622.112200] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 622.112200] env[66641]: value = "task-5145540" [ 622.112200] env[66641]: _type = "Task" [ 622.112200] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.112431] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60587abc-19ea-4f91-b656-490513917be2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.120454] env[66641]: DEBUG nova.network.neutron [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Updating instance_info_cache with network_info: [{"id": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "address": "fa:16:3e:1a:5e:cc", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap361c7a50-fe", "ovs_interfaceid": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 622.124800] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145540, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.126456] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 622.126456] env[66641]: value = "task-5145541" [ 622.126456] env[66641]: _type = "Task" [ 622.126456] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.135128] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145541, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.164798] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.165067] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 622.165258] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 622.273928] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 622.310183] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 622.310384] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 622.310840] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 622.310840] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 622.310946] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 622.311127] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 622.311359] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 622.311412] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 622.311574] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 622.311729] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 622.311932] env[66641]: DEBUG nova.virt.hardware [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 622.312697] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 622.315320] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31bfaa1a-2f1a-4143-98f3-4dc3b048af53 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.329963] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457c7d04-c2d1-47e2-9260-5f5d86bec84a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.378462] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Successfully created port: 71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 622.627089] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145540, 'name': ReconfigVM_Task, 'duration_secs': 0.325436} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 622.627411] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Releasing lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 622.627667] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Instance network_info: |[{"id": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "address": "fa:16:3e:1a:5e:cc", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap361c7a50-fe", "ovs_interfaceid": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 622.628149] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 4ebc0f2b-bb79-4941-b8f2-082560c80f8b/4ebc0f2b-bb79-4941-b8f2-082560c80f8b.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 622.629779] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:5e:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '361c7a50-fe6b-4975-bf4e-5bceab411bc7', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 622.641808] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Creating folder: Project (09cb6ffdd65e4f579dabc938cadc366d). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.650502] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98706a68-46a0-4541-b477-22f1f346806f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.650502] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2855a175-dfe7-48d1-827a-15591191c8b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.659580] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145541, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.663105] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 622.663105] env[66641]: value = "task-5145542" [ 622.663105] env[66641]: _type = "Task" [ 622.663105] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.669162] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 622.669468] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 622.674959] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 622.678682] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Created folder: Project (09cb6ffdd65e4f579dabc938cadc366d) in parent group-v1000566. [ 622.678937] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Creating folder: Instances. Parent ref: group-v1000573. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 622.679436] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ccc45ee3-7e80-4ba3-931f-4acbe99bd66e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.684897] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145542, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 622.693904] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Created folder: Instances in parent group-v1000573. [ 622.694193] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 622.694407] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 622.694625] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9d2fdaf4-a992-4e41-afb7-12cbb849877e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.716826] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 622.716826] env[66641]: value = "task-5145545" [ 622.716826] env[66641]: _type = "Task" [ 622.716826] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 622.821566] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 47aaed25-542b-4ceb-9adf-6a4953c8c95d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 622.822028] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 622.822198] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=100GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '5', 'num_vm_building': '5', 'num_task_spawning': '4', 'num_os_type_None': '5', 'num_proj_37c028dc169a444fa76652ff45efc917': '1', 'io_workload': '5', 'num_proj_09cb6ffdd65e4f579dabc938cadc366d': '1', 'num_proj_c00dc422cedf4cec85b175435c6d19fc': '1', 'num_proj_82383b2f8cf2490ebdf92495d9dc20e6': '1', 'num_task_None': '1', 'num_proj_417e1e40aca447229001fa725c82e9ca': '1'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 622.954481] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573db14e-2b98-4116-a062-d9ed881e594a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.963093] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40a38205-7f7b-4c23-85be-9c79d255fca8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.994204] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc8e5330-8f24-4a93-9dec-0fc39cbd9db3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.002746] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7e0668-86ff-4ac0-9b19-09e01e994fb2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.019448] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.138618] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145541, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.175970] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145542, 'name': Rename_Task, 'duration_secs': 0.19905} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.175970] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 623.175970] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-63ae2035-0718-4bca-9d70-b4aaa0224c5a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.182848] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 623.182848] env[66641]: value = "task-5145546" [ 623.182848] env[66641]: _type = "Task" [ 623.182848] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.194395] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145546, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.227796] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145545, 'name': CreateVM_Task, 'duration_secs': 0.405922} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.232940] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 623.234254] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 623.234590] env[66641]: WARNING openstack [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 623.239819] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.240558] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.241535] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 623.241535] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-36285d6b-d0bf-403a-ba7f-a690cf4cf411 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.247513] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 623.247513] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5251143d-6522-df03-b055-75e5fbf86f89" [ 623.247513] env[66641]: _type = "Task" [ 623.247513] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.257244] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5251143d-6522-df03-b055-75e5fbf86f89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.523629] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 623.586236] env[66641]: DEBUG nova.compute.manager [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Received event network-vif-plugged-361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 623.586236] env[66641]: DEBUG oslo_concurrency.lockutils [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] Acquiring lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 623.586456] env[66641]: DEBUG oslo_concurrency.lockutils [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 623.586456] env[66641]: DEBUG oslo_concurrency.lockutils [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 623.586818] env[66641]: DEBUG nova.compute.manager [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] No waiting events found dispatching network-vif-plugged-361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 623.586818] env[66641]: WARNING nova.compute.manager [req-62ff14be-485d-47a6-b7bf-bd66a6892e9c req-fd33203b-ecdf-43d5-b772-fc1c4f34d7ef service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Received unexpected event network-vif-plugged-361c7a50-fe6b-4975-bf4e-5bceab411bc7 for instance with vm_state building and task_state spawning. [ 623.639728] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145541, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.693944] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145546, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 623.759023] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5251143d-6522-df03-b055-75e5fbf86f89, 'name': SearchDatastore_Task, 'duration_secs': 0.029956} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 623.759023] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 623.759023] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 623.759023] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.759316] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 623.759316] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 623.759316] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca984b52-bfd5-4439-a94b-2c27ce8c1f0f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.768397] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 623.768654] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 623.769420] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd7b3502-1326-4054-b6b0-152979a63a13 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.775513] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 623.775513] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c44443-41c6-33e2-3135-26ab02a8b504" [ 623.775513] env[66641]: _type = "Task" [ 623.775513] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 623.784346] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c44443-41c6-33e2-3135-26ab02a8b504, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.028603] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 624.028807] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.289s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 624.029079] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.515s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 624.030650] env[66641]: INFO nova.compute.claims [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.140540] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145541, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.527123} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.140797] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 25203771-cf4d-4fc1-a1bf-f2a0200296b7/25203771-cf4d-4fc1-a1bf-f2a0200296b7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 624.141013] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 624.141264] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c6725bae-0290-44b7-8bbc-656dd7a95e25 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.149705] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 624.149705] env[66641]: value = "task-5145547" [ 624.149705] env[66641]: _type = "Task" [ 624.149705] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.161612] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145547, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.193975] env[66641]: DEBUG oslo_vmware.api [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145546, 'name': PowerOnVM_Task, 'duration_secs': 0.571334} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.196492] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Successfully updated port: 71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 624.197695] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 624.198917] env[66641]: INFO nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Took 9.25 seconds to spawn the instance on the hypervisor. [ 624.198917] env[66641]: DEBUG nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 624.202762] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96bd81ad-624d-48ee-8f60-712579e92558 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.216162] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 624.216559] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 624.287152] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c44443-41c6-33e2-3135-26ab02a8b504, 'name': SearchDatastore_Task, 'duration_secs': 0.009095} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.287977] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11e998c3-f6c4-4098-a652-132f1144ee87 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.294358] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 624.294358] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52713051-7e3c-589c-498d-a55a7a955195" [ 624.294358] env[66641]: _type = "Task" [ 624.294358] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.308725] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52713051-7e3c-589c-498d-a55a7a955195, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.663478] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145547, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067495} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.663921] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 624.664861] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72aebe73-8337-4f63-8261-4934c7ed091e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.694876] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Reconfiguring VM instance instance-00000001 to attach disk [datastore2] 25203771-cf4d-4fc1-a1bf-f2a0200296b7/25203771-cf4d-4fc1-a1bf-f2a0200296b7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 624.695354] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e5b1ec4-367c-4e63-9545-e8c22c613ef8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.712062] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.712362] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 624.712716] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 624.720870] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 624.720870] env[66641]: value = "task-5145548" [ 624.720870] env[66641]: _type = "Task" [ 624.720870] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.734894] env[66641]: INFO nova.compute.manager [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Took 16.19 seconds to build instance. [ 624.745016] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145548, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 624.806755] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52713051-7e3c-589c-498d-a55a7a955195, 'name': SearchDatastore_Task, 'duration_secs': 0.009464} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 624.809017] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 624.809017] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] ccf46c5c-1c79-4672-ad9f-ea61042097d5/ccf46c5c-1c79-4672-ad9f-ea61042097d5.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 624.809017] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9c93ce7a-2659-44ea-b613-f72a627bf878 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.816771] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 624.816771] env[66641]: value = "task-5145549" [ 624.816771] env[66641]: _type = "Task" [ 624.816771] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 624.828463] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145549, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.136175] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.136613] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.216945] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.217503] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.224992] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 625.230520] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7172616e-dfda-4df9-adcb-47f09d813d87 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.247605] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28104fb-a19b-4ac6-81bc-755382dfb944 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.251392] env[66641]: DEBUG oslo_concurrency.lockutils [None req-41460eaf-2162-4968-877e-09324fc9410a tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.721s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 625.252445] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145548, 'name': ReconfigVM_Task, 'duration_secs': 0.334581} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.252587] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Reconfigured VM instance instance-00000001 to attach disk [datastore2] 25203771-cf4d-4fc1-a1bf-f2a0200296b7/25203771-cf4d-4fc1-a1bf-f2a0200296b7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 625.253678] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-89d0289f-a713-46d5-897b-5febf67ec355 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.284925] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db1d029c-8073-4c98-abe9-09a6f4b75ed6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.288564] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 625.288564] env[66641]: value = "task-5145550" [ 625.288564] env[66641]: _type = "Task" [ 625.288564] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.297168] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e703d79-3d37-4395-aef0-692abb17ea8d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.305259] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145550, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.316896] env[66641]: DEBUG nova.compute.provider_tree [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.328364] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145549, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.746493] env[66641]: DEBUG nova.network.neutron [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Updating instance_info_cache with network_info: [{"id": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "address": "fa:16:3e:3c:ab:ed", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88f6140a-20", "ovs_interfaceid": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 625.761887] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 625.761887] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 625.804384] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145550, 'name': Rename_Task, 'duration_secs': 0.215594} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.804384] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 625.804384] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c411364a-2f59-45da-9400-aa35595aac4d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.812027] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 625.812027] env[66641]: value = "task-5145551" [ 625.812027] env[66641]: _type = "Task" [ 625.812027] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.824049] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145551, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 625.829671] env[66641]: DEBUG nova.scheduler.client.report [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 625.848677] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145549, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.537595} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 625.849277] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] ccf46c5c-1c79-4672-ad9f-ea61042097d5/ccf46c5c-1c79-4672-ad9f-ea61042097d5.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 625.849535] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 625.849801] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d2d420c5-61ae-473b-8f21-cfa6bd0c55aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.862497] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 625.862497] env[66641]: value = "task-5145552" [ 625.862497] env[66641]: _type = "Task" [ 625.862497] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 625.874572] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145552, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.254025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 626.254025] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Instance network_info: |[{"id": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "address": "fa:16:3e:3c:ab:ed", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88f6140a-20", "ovs_interfaceid": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 626.254328] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:ab:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '88f6140a-2061-43ef-9eed-eea8ab8878c5', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 626.272243] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Creating folder: Project (82383b2f8cf2490ebdf92495d9dc20e6). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.272857] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cfac7cc5-fad0-4f66-8a7d-0a54da850bc3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.286822] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Created folder: Project (82383b2f8cf2490ebdf92495d9dc20e6) in parent group-v1000566. [ 626.286822] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Creating folder: Instances. Parent ref: group-v1000576. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 626.286822] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-6a60806f-6c4d-4ccd-bfb4-f8885e80e9f7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.302081] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Created folder: Instances in parent group-v1000576. [ 626.302081] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 626.302081] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 626.302081] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-023b3a84-8bf8-4dd7-b8bf-17b88ce7ed13 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.331519] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.332903] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.343929] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 626.344755] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 626.358594] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.821s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 626.361183] env[66641]: INFO nova.compute.claims [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.364452] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 626.364452] env[66641]: value = "task-5145555" [ 626.364452] env[66641]: _type = "Task" [ 626.364452] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.365319] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145551, 'name': PowerOnVM_Task} progress is 87%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.386492] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145555, 'name': CreateVM_Task} progress is 15%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.386492] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145552, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155221} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 626.386722] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 626.387753] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c6b0e7-c3a3-4b69-aa18-ddf934a161a1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.416949] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] ccf46c5c-1c79-4672-ad9f-ea61042097d5/ccf46c5c-1c79-4672-ad9f-ea61042097d5.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 626.418412] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fd61a258-041e-4314-ac3a-239bb0d38716 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.445932] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 626.445932] env[66641]: value = "task-5145556" [ 626.445932] env[66641]: _type = "Task" [ 626.445932] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 626.458238] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.659853] env[66641]: DEBUG nova.network.neutron [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Updating instance_info_cache with network_info: [{"id": "71cb13b6-7303-4801-a446-6913f2523c32", "address": "fa:16:3e:95:b9:5f", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71cb13b6-73", "ovs_interfaceid": "71cb13b6-7303-4801-a446-6913f2523c32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 626.767877] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Received event network-changed-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 626.768308] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Refreshing instance network info cache due to event network-changed-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 626.769373] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Acquiring lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 626.769557] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Acquired lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 626.769737] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Refreshing network info cache for port 0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 626.832656] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145551, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.869392] env[66641]: DEBUG nova.compute.utils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 626.871904] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 626.872134] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 626.872497] env[66641]: WARNING neutronclient.v2_0.client [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.873161] env[66641]: WARNING neutronclient.v2_0.client [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 626.873770] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 626.874268] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 626.895831] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145555, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.962273] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145556, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 626.971073] env[66641]: DEBUG nova.policy [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfb62f2b801a4c7a899ac65c05c382f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6624f269bb8840e98c10259f006ce962', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 627.164396] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.164488] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Instance network_info: |[{"id": "71cb13b6-7303-4801-a446-6913f2523c32", "address": "fa:16:3e:95:b9:5f", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71cb13b6-73", "ovs_interfaceid": "71cb13b6-7303-4801-a446-6913f2523c32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 627.165249] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:95:b9:5f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b5c60ce-845e-4506-bc10-348461fece6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '71cb13b6-7303-4801-a446-6913f2523c32', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 627.178836] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Creating folder: Project (417e1e40aca447229001fa725c82e9ca). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 627.181071] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a13e6011-f377-4f6c-ad4f-cc10a4dc93c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.194658] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Created folder: Project (417e1e40aca447229001fa725c82e9ca) in parent group-v1000566. [ 627.194871] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Creating folder: Instances. Parent ref: group-v1000579. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 627.195154] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0042f260-9a3e-432c-82c1-6b60027ccd51 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.207399] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Created folder: Instances in parent group-v1000579. [ 627.208370] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 627.208370] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 627.208370] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-957c0391-e08c-4d08-b495-30de4fb4d58d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.231045] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 627.231045] env[66641]: value = "task-5145559" [ 627.231045] env[66641]: _type = "Task" [ 627.231045] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.245253] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145559, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.273694] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.274441] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.341333] env[66641]: DEBUG oslo_vmware.api [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145551, 'name': PowerOnVM_Task, 'duration_secs': 1.049862} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.341648] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 627.341867] env[66641]: INFO nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Took 14.64 seconds to spawn the instance on the hypervisor. [ 627.342069] env[66641]: DEBUG nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 627.342930] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a190a0a-95e8-4b53-9ed6-73f4f89c4dd8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.373268] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 627.383810] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Successfully created port: 5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 627.399813] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145555, 'name': CreateVM_Task, 'duration_secs': 0.58927} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.399986] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.401920] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.401920] env[66641]: WARNING openstack [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.409962] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.410129] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.410460] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 627.410777] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6f4543d-4579-4124-b985-62fb90e361e3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.423437] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 627.423437] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ec5dad-80c8-ccf3-1721-83d445b08ee9" [ 627.423437] env[66641]: _type = "Task" [ 627.423437] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.435813] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ec5dad-80c8-ccf3-1721-83d445b08ee9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.472574] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145556, 'name': ReconfigVM_Task, 'duration_secs': 0.748959} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.472574] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Reconfigured VM instance instance-00000002 to attach disk [datastore2] ccf46c5c-1c79-4672-ad9f-ea61042097d5/ccf46c5c-1c79-4672-ad9f-ea61042097d5.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 627.472574] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-33a655f9-bdd3-4c94-93fb-5db9657cd60e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.481892] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 627.481892] env[66641]: value = "task-5145560" [ 627.481892] env[66641]: _type = "Task" [ 627.481892] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.492927] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145560, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.581473] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f352deb-6c9d-4c7c-8ea1-195bdfd751d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.591266] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09525d5-7493-4f95-be70-4239014b8c7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.637741] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acc07ab3-35cc-45c2-a31f-737e44686d6b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.648798] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c59bd5e-b039-4cba-b7f8-64755c1c982c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.665678] env[66641]: DEBUG nova.compute.provider_tree [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.688729] env[66641]: DEBUG nova.compute.manager [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Received event network-changed-361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 627.689024] env[66641]: DEBUG nova.compute.manager [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Refreshing instance network info cache due to event network-changed-361c7a50-fe6b-4975-bf4e-5bceab411bc7. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 627.689226] env[66641]: DEBUG oslo_concurrency.lockutils [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Acquiring lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.689432] env[66641]: DEBUG oslo_concurrency.lockutils [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Acquired lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.689602] env[66641]: DEBUG nova.network.neutron [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Refreshing network info cache for port 361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 627.748012] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145559, 'name': CreateVM_Task, 'duration_secs': 0.453205} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.748012] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 627.748536] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.748909] env[66641]: WARNING openstack [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.754103] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.851465] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 627.851849] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 627.864456] env[66641]: INFO nova.compute.manager [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Took 19.52 seconds to build instance. [ 627.872250] env[66641]: DEBUG nova.compute.manager [None req-574bb9c4-5669-45b9-a586-72b964c391b2 tempest-ServerDiagnosticsV248Test-2136313891 tempest-ServerDiagnosticsV248Test-2136313891-project-admin] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 627.874236] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac76ca2-5c86-41ae-b2e5-bde4540f5fe0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.883032] env[66641]: INFO nova.compute.manager [None req-574bb9c4-5669-45b9-a586-72b964c391b2 tempest-ServerDiagnosticsV248Test-2136313891 tempest-ServerDiagnosticsV248Test-2136313891-project-admin] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Retrieving diagnostics [ 627.883885] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516627a6-c7f8-4061-8feb-edeb9e2cb5a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.936254] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ec5dad-80c8-ccf3-1721-83d445b08ee9, 'name': SearchDatastore_Task, 'duration_secs': 0.017413} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.936433] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 627.936667] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 627.936895] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.937052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.937228] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 627.937496] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 627.937972] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 627.938087] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dcf9ec07-9728-4d3c-bcd0-4087bfb253f4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.939990] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ce38df9a-f228-48fe-8f19-94abac8e3bee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.945365] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 627.945365] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b04cd2-214e-4194-4b92-32d59cd68c65" [ 627.945365] env[66641]: _type = "Task" [ 627.945365] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.956349] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 627.956540] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 627.960244] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-778b7a73-cc95-4c1e-9ac1-914d46b893be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.963322] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b04cd2-214e-4194-4b92-32d59cd68c65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.967513] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 627.967513] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521130b6-73ec-191b-fbb8-f6e7236bc012" [ 627.967513] env[66641]: _type = "Task" [ 627.967513] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 627.976488] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521130b6-73ec-191b-fbb8-f6e7236bc012, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 627.992828] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145560, 'name': Rename_Task, 'duration_secs': 0.197207} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 627.992828] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 627.993309] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e3ba4ed6-8a13-4838-98f4-8173a840deaa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.001835] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 628.001835] env[66641]: value = "task-5145561" [ 628.001835] env[66641]: _type = "Task" [ 628.001835] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.010810] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145561, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.169634] env[66641]: DEBUG nova.scheduler.client.report [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 628.190263] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 628.190263] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 628.193111] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 628.193471] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 628.366572] env[66641]: DEBUG oslo_concurrency.lockutils [None req-500f75f4-557c-4413-85f7-982b19cbe789 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.035s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.422540] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 628.453439] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 628.453439] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 628.453647] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 628.453800] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 628.453931] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 628.454799] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 628.454799] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.455191] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 628.455514] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 628.455617] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 628.455853] env[66641]: DEBUG nova.virt.hardware [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 628.457327] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf23700d-507a-4c97-8700-678622bbbd4f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.470978] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e1cd86-eab2-4916-ba0a-0591ba1796b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.475676] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b04cd2-214e-4194-4b92-32d59cd68c65, 'name': SearchDatastore_Task, 'duration_secs': 0.02634} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.478987] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.479278] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 628.480031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.486397] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521130b6-73ec-191b-fbb8-f6e7236bc012, 'name': SearchDatastore_Task, 'duration_secs': 0.02679} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.496325] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0b1f657a-2a0f-43b3-8e03-5445e7f8f3a3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.507308] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 628.507308] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5271df16-bf78-48b9-bb64-abba18f7b54e" [ 628.507308] env[66641]: _type = "Task" [ 628.507308] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.516854] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145561, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.523038] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5271df16-bf78-48b9-bb64-abba18f7b54e, 'name': SearchDatastore_Task, 'duration_secs': 0.015104} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 628.523349] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 628.523622] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c/420544a8-1a02-42d5-8a9a-e3e8b6a11a0c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 628.523947] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 628.524198] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 628.524424] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbf6a279-7610-4be7-aa4f-cdbd32946ab0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.526533] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0c672d13-7e0e-4f75-b35e-c4656aa70ab7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.534612] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 628.534612] env[66641]: value = "task-5145562" [ 628.534612] env[66641]: _type = "Task" [ 628.534612] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.535954] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 628.536059] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 628.539809] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b0951a-f22d-4560-987f-59909e37c297 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.548162] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145562, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.549754] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 628.549754] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521b3c04-84bd-85af-26d4-ababa0a1a63c" [ 628.549754] env[66641]: _type = "Task" [ 628.549754] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 628.559309] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521b3c04-84bd-85af-26d4-ababa0a1a63c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 628.675477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 628.676120] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 628.691865] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 629.012741] env[66641]: DEBUG oslo_vmware.api [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145561, 'name': PowerOnVM_Task, 'duration_secs': 0.710611} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.013157] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 629.013515] env[66641]: INFO nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Took 11.75 seconds to spawn the instance on the hypervisor. [ 629.013515] env[66641]: DEBUG nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 629.014541] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a664495-ba46-4435-80ed-df9abe3e17b4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.049030] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145562, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.061086] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521b3c04-84bd-85af-26d4-ababa0a1a63c, 'name': SearchDatastore_Task, 'duration_secs': 0.012068} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.062106] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e4e1b84-bb2b-4215-8307-bb0a60efd722 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.069561] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 629.069561] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f890ec-670f-4d80-10d1-d04d63dae9a4" [ 629.069561] env[66641]: _type = "Task" [ 629.069561] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.081190] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f890ec-670f-4d80-10d1-d04d63dae9a4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.089424] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Successfully updated port: 5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 629.124913] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.125420] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.182271] env[66641]: DEBUG nova.compute.utils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 629.184481] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 629.184765] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 629.185227] env[66641]: WARNING neutronclient.v2_0.client [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.185651] env[66641]: WARNING neutronclient.v2_0.client [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 629.186489] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 629.186987] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 629.226142] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 629.226526] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 629.229256] env[66641]: INFO nova.compute.claims [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 629.534488] env[66641]: INFO nova.compute.manager [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Took 20.97 seconds to build instance. [ 629.552609] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145562, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.878051} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.552873] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c/420544a8-1a02-42d5-8a9a-e3e8b6a11a0c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 629.553099] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 629.553367] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff204532-3fb2-4b46-9cc0-f9ea6fe778f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.561750] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 629.561750] env[66641]: value = "task-5145563" [ 629.561750] env[66641]: _type = "Task" [ 629.561750] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.574917] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145563, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.588171] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f890ec-670f-4d80-10d1-d04d63dae9a4, 'name': SearchDatastore_Task, 'duration_secs': 0.073269} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 629.588475] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 629.588955] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3/d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 629.589071] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-338ea960-4617-4551-b34f-c4992eef2176 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.593884] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.593884] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 629.593884] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 629.600437] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 629.600437] env[66641]: value = "task-5145564" [ 629.600437] env[66641]: _type = "Task" [ 629.600437] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 629.613110] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145564, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 629.699281] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 629.814654] env[66641]: DEBUG nova.policy [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8085d25a33c94c5092e444eab1d69ec0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6911bbec11fa43a4a2db344655d00d32', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 630.040089] env[66641]: DEBUG oslo_concurrency.lockutils [None req-09400f81-7005-4290-b586-46a513327496 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.488s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 630.072062] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145563, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078104} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 630.072370] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 630.073183] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ddc077a-dbc2-4d65-843e-fc8efaf1dbc5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.098740] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Reconfiguring VM instance instance-00000004 to attach disk [datastore2] 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c/420544a8-1a02-42d5-8a9a-e3e8b6a11a0c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 630.098740] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.098740] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.108280] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 630.108985] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cce3c31-721b-48ee-b186-f4076e4c13ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.143239] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145564, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.144933] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 630.144933] env[66641]: value = "task-5145565" [ 630.144933] env[66641]: _type = "Task" [ 630.144933] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 630.157466] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.406399] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7196b991-61a0-4d23-8f8a-6dadeaabe9ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.416377] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4526bb62-af9e-4fe5-b724-7a96e3d493e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.464659] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 630.465094] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 630.473565] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052b7012-1ae9-46a2-af8c-38f3ea9a6a91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.483902] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bcb38e1-64ae-457a-b183-36a2da974bfa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.503837] env[66641]: DEBUG nova.compute.provider_tree [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.651097] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145564, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.662175] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 630.675779] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Updated VIF entry in instance network info cache for port 0c57e5de-8952-4f18-8e92-c4cc1fb19ef7. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 630.676224] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Updating instance_info_cache with network_info: [{"id": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "address": "fa:16:3e:d6:f8:82", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.76", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c57e5de-89", "ovs_interfaceid": "0c57e5de-8952-4f18-8e92-c4cc1fb19ef7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 630.716638] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 630.747965] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 630.748433] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 630.748433] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 630.748583] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 630.748672] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 630.748787] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 630.748987] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 630.749172] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 630.749368] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 630.749553] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 630.749719] env[66641]: DEBUG nova.virt.hardware [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 630.750661] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ee9fafb-1cec-4a33-b63c-6dbd018d069d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.759265] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d3ca9e1-8272-4df4-8245-22e5e18abe01 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.898170] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Successfully created port: 46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 631.008059] env[66641]: DEBUG nova.scheduler.client.report [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 631.144485] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145564, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.084664} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.144764] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3/d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 631.144934] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 631.145272] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5bf8852-f396-4603-8c90-134c3dba3439 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.155777] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 631.155777] env[66641]: value = "task-5145566" [ 631.155777] env[66641]: _type = "Task" [ 631.155777] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.165469] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.170520] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145566, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.180163] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Releasing lock "refresh_cache-25203771-cf4d-4fc1-a1bf-f2a0200296b7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 631.180163] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Received event network-vif-plugged-88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 631.180163] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Acquiring lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 631.180163] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 631.180163] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.180506] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] No waiting events found dispatching network-vif-plugged-88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 631.180506] env[66641]: WARNING nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Received unexpected event network-vif-plugged-88f6140a-2061-43ef-9eed-eea8ab8878c5 for instance with vm_state building and task_state spawning. [ 631.180621] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Received event network-changed-88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 631.180709] env[66641]: DEBUG nova.compute.manager [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Refreshing instance network info cache due to event network-changed-88f6140a-2061-43ef-9eed-eea8ab8878c5. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 631.180868] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Acquiring lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.181008] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Acquired lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 631.181161] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Refreshing network info cache for port 88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 631.249753] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.250165] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.281294] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.282628] env[66641]: WARNING openstack [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.515492] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.289s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 631.515822] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 631.661083] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 631.672136] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145566, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076945} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 631.673861] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 631.673861] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7064eb0b-a4a3-4d33-9b6d-92ebefbbfc9e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.689371] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 631.690275] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 631.707797] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Reconfiguring VM instance instance-00000005 to attach disk [datastore2] d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3/d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 631.708822] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-42560b23-6eb5-4f5a-95ce-e008243804fa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.734390] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 631.734390] env[66641]: value = "task-5145567" [ 631.734390] env[66641]: _type = "Task" [ 631.734390] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 631.744878] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145567, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.022397] env[66641]: DEBUG nova.compute.utils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 632.023896] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 632.024719] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 632.025349] env[66641]: WARNING neutronclient.v2_0.client [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.025714] env[66641]: WARNING neutronclient.v2_0.client [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 632.026488] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.026987] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.161643] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.246316] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145567, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.537453] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 632.642525] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 632.643318] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 632.663553] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145565, 'name': ReconfigVM_Task, 'duration_secs': 2.416691} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.664121] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Reconfigured VM instance instance-00000004 to attach disk [datastore2] 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c/420544a8-1a02-42d5-8a9a-e3e8b6a11a0c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.664962] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-279c6391-3cf9-4757-b035-9d816c50b52a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.674979] env[66641]: DEBUG nova.network.neutron [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Updated VIF entry in instance network info cache for port 361c7a50-fe6b-4975-bf4e-5bceab411bc7. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 632.674979] env[66641]: DEBUG nova.network.neutron [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Updating instance_info_cache with network_info: [{"id": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "address": "fa:16:3e:1a:5e:cc", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap361c7a50-fe", "ovs_interfaceid": "361c7a50-fe6b-4975-bf4e-5bceab411bc7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 632.677758] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 632.677758] env[66641]: value = "task-5145568" [ 632.677758] env[66641]: _type = "Task" [ 632.677758] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.690752] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145568, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.738989] env[66641]: DEBUG nova.policy [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '337d565387724ae2a92d2655d212150c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e25db00afd9044e3940b419d37751237', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 632.757127] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145567, 'name': ReconfigVM_Task, 'duration_secs': 0.785692} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 632.757435] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Reconfigured VM instance instance-00000005 to attach disk [datastore2] d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3/d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 632.758611] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0cd3809e-7770-498d-adc8-631e5852d6e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.767824] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 632.767824] env[66641]: value = "task-5145569" [ 632.767824] env[66641]: _type = "Task" [ 632.767824] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 632.777524] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145569, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 632.899353] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Successfully updated port: 46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 633.177033] env[66641]: DEBUG oslo_concurrency.lockutils [req-c4721075-ad05-4980-b92e-4a11612c88cd req-5956252d-e169-4ee0-9f5c-f11ccc2745bc service nova] Releasing lock "refresh_cache-ccf46c5c-1c79-4672-ad9f-ea61042097d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 633.192141] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145568, 'name': Rename_Task, 'duration_secs': 0.245807} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.192630] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.192903] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b88b33a0-38a2-4fad-926f-be5b84e108d4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.201120] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 633.201120] env[66641]: value = "task-5145570" [ 633.201120] env[66641]: _type = "Task" [ 633.201120] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.214107] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.240195] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Successfully created port: 3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 633.279416] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145569, 'name': Rename_Task, 'duration_secs': 0.190383} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 633.279773] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 633.280031] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c948165-262f-45ed-91d2-df236ad993e3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.287769] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 633.287769] env[66641]: value = "task-5145571" [ 633.287769] env[66641]: _type = "Task" [ 633.287769] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 633.298536] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145571, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.407244] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.407244] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquired lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 633.407244] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 633.415424] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.415827] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.505161] env[66641]: DEBUG nova.network.neutron [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 633.546484] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 633.579171] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 633.579473] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 633.579653] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 633.579842] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 633.579982] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 633.580296] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 633.580617] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 633.580776] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 633.580955] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 633.581132] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 633.581326] env[66641]: DEBUG nova.virt.hardware [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 633.582828] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5570c9ae-56b7-41a0-96f0-14cb0106bb2b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.591741] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac1e944-be08-45dc-87dd-dd520a48bc63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.693677] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 633.713175] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145570, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.800870] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145571, 'name': PowerOnVM_Task} progress is 1%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 633.910208] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 633.910696] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 633.917992] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 634.008937] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 634.009470] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Instance network_info: |[{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 634.009999] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8a:b6:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5ebec271-9d57-41f2-ba8c-59edb82c50db', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 634.018895] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating folder: Project (6624f269bb8840e98c10259f006ce962). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 634.019842] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e9806d54-f9e3-4806-8cb2-9f340b224b11 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.033698] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Created folder: Project (6624f269bb8840e98c10259f006ce962) in parent group-v1000566. [ 634.034098] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating folder: Instances. Parent ref: group-v1000582. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 634.034442] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-930ebc7d-e6aa-476a-8e01-46e173f91442 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.048099] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Created folder: Instances in parent group-v1000582. [ 634.048384] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 634.048600] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 634.048820] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-60b28761-e097-43e0-beaa-80403cb84dbf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.074540] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 634.074540] env[66641]: value = "task-5145574" [ 634.074540] env[66641]: _type = "Task" [ 634.074540] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 634.083658] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145574, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.216327] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145570, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.300025] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145571, 'name': PowerOnVM_Task} progress is 86%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.588487] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145574, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 634.713348] env[66641]: DEBUG oslo_vmware.api [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145570, 'name': PowerOnVM_Task, 'duration_secs': 1.035329} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.713613] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.713824] env[66641]: INFO nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Took 15.21 seconds to spawn the instance on the hypervisor. [ 634.713988] env[66641]: DEBUG nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.714834] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02b8ac1c-9889-4d45-b560-9dff659723d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.810772] env[66641]: DEBUG oslo_vmware.api [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145571, 'name': PowerOnVM_Task, 'duration_secs': 1.059833} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 634.811158] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 634.811298] env[66641]: INFO nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Took 12.54 seconds to spawn the instance on the hypervisor. [ 634.811473] env[66641]: DEBUG nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 634.812303] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385de7da-5b96-4aa0-859b-c150764dde47 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.086597] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145574, 'name': CreateVM_Task, 'duration_secs': 0.567281} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 635.086852] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 635.087687] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.088158] env[66641]: WARNING openstack [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.094556] env[66641]: DEBUG oslo_vmware.service [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d18138-4288-4709-b6ae-b66ae07b94e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.101850] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.102429] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.102653] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 635.102987] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bfb86f1-2b27-4899-969a-6ef8e74adc80 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.109018] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 635.109018] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a821e5-1667-2c4a-a79f-d8a7127d6ccc" [ 635.109018] env[66641]: _type = "Task" [ 635.109018] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.118206] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a821e5-1667-2c4a-a79f-d8a7127d6ccc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.237398] env[66641]: INFO nova.compute.manager [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Took 24.44 seconds to build instance. [ 635.328661] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.328797] env[66641]: WARNING openstack [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.347553] env[66641]: INFO nova.compute.manager [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Took 22.68 seconds to build instance. [ 635.573193] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Successfully updated port: 3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 635.586961] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 635.589374] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 635.622238] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 635.622703] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 635.622963] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.622963] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.623220] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 635.623567] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f69cc2dc-3b48-47f7-941e-fa52e357226a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.644937] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 635.645160] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 635.646268] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a62a4f-028e-45b5-b604-d60cefe1570a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.656757] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f3ca53d-71f4-4eb6-a916-922a3feb0296 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.663261] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 635.663261] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52be935f-1f67-c273-46f2-b23c865d1e53" [ 635.663261] env[66641]: _type = "Task" [ 635.663261] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 635.674835] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52be935f-1f67-c273-46f2-b23c865d1e53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 635.738573] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5b1062bf-0463-4593-b394-ff9f5e6ffab4 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.957s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.851079] env[66641]: DEBUG oslo_concurrency.lockutils [None req-747f5f24-21c6-4f9f-b4bd-5bac155eba42 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.196s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.859414] env[66641]: DEBUG nova.compute.manager [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Received event network-vif-plugged-71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 635.859736] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Acquiring lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 635.863111] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 635.863111] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 635.863111] env[66641]: DEBUG nova.compute.manager [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] No waiting events found dispatching network-vif-plugged-71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 635.863111] env[66641]: WARNING nova.compute.manager [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Received unexpected event network-vif-plugged-71cb13b6-7303-4801-a446-6913f2523c32 for instance with vm_state active and task_state None. [ 635.863111] env[66641]: DEBUG nova.compute.manager [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Received event network-changed-71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 635.863528] env[66641]: DEBUG nova.compute.manager [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Refreshing instance network info cache due to event network-changed-71cb13b6-7303-4801-a446-6913f2523c32. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 635.863528] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Acquiring lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.863528] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Acquired lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 635.863528] env[66641]: DEBUG nova.network.neutron [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Refreshing network info cache for port 71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 636.014982] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Updated VIF entry in instance network info cache for port 88f6140a-2061-43ef-9eed-eea8ab8878c5. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 636.014982] env[66641]: DEBUG nova.network.neutron [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Updating instance_info_cache with network_info: [{"id": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "address": "fa:16:3e:3c:ab:ed", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.231", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap88f6140a-20", "ovs_interfaceid": "88f6140a-2061-43ef-9eed-eea8ab8878c5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 636.076697] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.077564] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquired lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.077564] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 636.138289] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.138689] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.175873] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Preparing fetch location {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 636.175933] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating directory with path [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 636.176192] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-204d5904-f886-4f85-b1c5-12450ae695a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.207067] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Created directory with path [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 636.207067] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Fetch image to [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 636.207067] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Downloading image file data cd910b37-6707-4868-b172-79fffc590a51 to [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk on the data store datastore1 {{(pid=66641) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 636.207067] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c14a39cb-cc27-457c-a2d2-fbdc4f21bafd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.213139] env[66641]: DEBUG nova.compute.manager [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Received event network-vif-plugged-5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 636.213480] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 636.215289] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 636.215658] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 636.215829] env[66641]: DEBUG nova.compute.manager [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] No waiting events found dispatching network-vif-plugged-5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 636.215860] env[66641]: WARNING nova.compute.manager [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Received unexpected event network-vif-plugged-5ebec271-9d57-41f2-ba8c-59edb82c50db for instance with vm_state building and task_state spawning. [ 636.216251] env[66641]: DEBUG nova.compute.manager [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Received event network-changed-5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 636.216251] env[66641]: DEBUG nova.compute.manager [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Refreshing instance network info cache due to event network-changed-5ebec271-9d57-41f2-ba8c-59edb82c50db. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 636.216459] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Acquiring lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.216492] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Acquired lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 636.216774] env[66641]: DEBUG nova.network.neutron [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Refreshing network info cache for port 5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 636.225696] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aed2ced-f11b-48a4-9f61-d0552607fe0f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.237346] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7476db-6194-482b-aad8-c33f635794b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.280203] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff34064-88df-40ba-b35e-501834dfa748 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.287552] env[66641]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-fb34455b-8251-44f6-a7a9-34ca24d7a504 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.322692] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Downloading image file data cd910b37-6707-4868-b172-79fffc590a51 to the data store datastore1 {{(pid=66641) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 636.365430] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.366021] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.390885] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=66641) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 636.518919] env[66641]: DEBUG oslo_concurrency.lockutils [req-8a821a12-b467-4b24-9a5d-7f8d57015853 req-f2893c6b-435e-4a54-a725-5b8c25d1594c service nova] Releasing lock "refresh_cache-420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 636.582306] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.582693] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.589875] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 636.722298] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 636.722689] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 636.797998] env[66641]: DEBUG nova.network.neutron [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Updating instance_info_cache with network_info: [{"id": "46c2027b-2796-42b8-9167-719bceab5640", "address": "fa:16:3e:d5:fb:69", "network": {"id": "7929b0b6-79ef-4cfa-a453-26990f6083fa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-826902442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6911bbec11fa43a4a2db344655d00d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46c2027b-27", "ovs_interfaceid": "46c2027b-2796-42b8-9167-719bceab5640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 637.275030] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.275642] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.300556] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Releasing lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 637.300955] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Instance network_info: |[{"id": "46c2027b-2796-42b8-9167-719bceab5640", "address": "fa:16:3e:d5:fb:69", "network": {"id": "7929b0b6-79ef-4cfa-a453-26990f6083fa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-826902442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6911bbec11fa43a4a2db344655d00d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46c2027b-27", "ovs_interfaceid": "46c2027b-2796-42b8-9167-719bceab5640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 637.301471] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:fb:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b9aabc7c-0f6c-42eb-bd27-493a1496c0c8', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46c2027b-2796-42b8-9167-719bceab5640', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 637.309846] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Creating folder: Project (6911bbec11fa43a4a2db344655d00d32). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.310270] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ad4a8a87-24d1-4510-94f8-09b32d367ba4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.322450] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Completed reading data from the image iterator. {{(pid=66641) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 637.322717] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 637.324574] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Created folder: Project (6911bbec11fa43a4a2db344655d00d32) in parent group-v1000566. [ 637.324784] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Creating folder: Instances. Parent ref: group-v1000585. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 637.325032] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48e8de46-6a22-4504-97a9-1fedf12af01b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.336600] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Created folder: Instances in parent group-v1000585. [ 637.337080] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 637.337566] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 637.337849] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ac43a704-d1eb-4c79-9ff0-48185ced8487 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.366145] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 637.366145] env[66641]: value = "task-5145577" [ 637.366145] env[66641]: _type = "Task" [ 637.366145] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.376985] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145577, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.382472] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Downloaded image file data cd910b37-6707-4868-b172-79fffc590a51 to vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk on the data store datastore1 {{(pid=66641) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 637.384275] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Caching image {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 637.384275] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copying Virtual Disk [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk to [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 637.384666] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-05f5bf7f-6054-471d-be80-683c43e45bd3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.395475] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 637.395475] env[66641]: value = "task-5145578" [ 637.395475] env[66641]: _type = "Task" [ 637.395475] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 637.405451] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.407034] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.407466] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.877965] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145577, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.906694] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145578, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 637.908958] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.909487] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 637.966911] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 637.969179] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.018857] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.018857] env[66641]: WARNING openstack [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.127108] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.129071] env[66641]: WARNING openstack [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.162948] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updating instance_info_cache with network_info: [{"id": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "address": "fa:16:3e:a2:b8:43", "network": {"id": "b6204f09-6035-4360-9da5-498dda901ff2", "bridge": "br-int", "label": "tempest-ServersTestJSON-324682247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25db00afd9044e3940b419d37751237", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e466382-fc", "ovs_interfaceid": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.287648] env[66641]: DEBUG nova.network.neutron [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Updated VIF entry in instance network info cache for port 71cb13b6-7303-4801-a446-6913f2523c32. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 638.288427] env[66641]: DEBUG nova.network.neutron [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Updating instance_info_cache with network_info: [{"id": "71cb13b6-7303-4801-a446-6913f2523c32", "address": "fa:16:3e:95:b9:5f", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap71cb13b6-73", "ovs_interfaceid": "71cb13b6-7303-4801-a446-6913f2523c32", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.325595] env[66641]: DEBUG nova.network.neutron [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updated VIF entry in instance network info cache for port 5ebec271-9d57-41f2-ba8c-59edb82c50db. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 638.325935] env[66641]: DEBUG nova.network.neutron [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 638.380818] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145577, 'name': CreateVM_Task, 'duration_secs': 0.707521} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 638.380818] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 638.382236] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 638.382708] env[66641]: WARNING openstack [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 638.390243] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.390616] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 638.391371] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 638.391371] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5c3821c-5305-4749-8bc3-9febbd058eff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.399926] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 638.399926] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f48786-3823-e627-77cd-54bd53ed5282" [ 638.399926] env[66641]: _type = "Task" [ 638.399926] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.414290] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145578, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.417882] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f48786-3823-e627-77cd-54bd53ed5282, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.600019] env[66641]: DEBUG nova.compute.manager [None req-a38b6fd1-4a91-498e-b446-b5f72c39d065 tempest-ServerDiagnosticsTest-1454730316 tempest-ServerDiagnosticsTest-1454730316-project-admin] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 638.601744] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5090c221-9ebc-4fbd-90ae-27c5d6d2ef3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.610321] env[66641]: INFO nova.compute.manager [None req-a38b6fd1-4a91-498e-b446-b5f72c39d065 tempest-ServerDiagnosticsTest-1454730316 tempest-ServerDiagnosticsTest-1454730316-project-admin] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Retrieving diagnostics [ 638.611209] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f614f2d0-fe8c-49d3-a9f9-0ae07a363c4c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.665823] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Releasing lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.666215] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance network_info: |[{"id": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "address": "fa:16:3e:a2:b8:43", "network": {"id": "b6204f09-6035-4360-9da5-498dda901ff2", "bridge": "br-int", "label": "tempest-ServersTestJSON-324682247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25db00afd9044e3940b419d37751237", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e466382-fc", "ovs_interfaceid": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 638.666676] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:b8:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '47ca1ce6-8148-48d5-bcfe-89e39b73914e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e466382-fc5f-4ac9-90dc-05e1d9f93c48', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 638.674709] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Creating folder: Project (e25db00afd9044e3940b419d37751237). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 638.674992] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4907ae12-aabb-4709-b036-ee69c5e21357 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.687231] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Created folder: Project (e25db00afd9044e3940b419d37751237) in parent group-v1000566. [ 638.688333] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Creating folder: Instances. Parent ref: group-v1000588. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 638.688333] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f496c4f-5004-46d6-947a-9a13011e295b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.700286] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Created folder: Instances in parent group-v1000588. [ 638.700286] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 638.700286] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 638.700286] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-548e57e6-5951-4fc6-878a-59bd0e041eba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.722063] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 638.722063] env[66641]: value = "task-5145581" [ 638.722063] env[66641]: _type = "Task" [ 638.722063] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 638.731026] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145581, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.790726] env[66641]: DEBUG oslo_concurrency.lockutils [req-2d69ef13-1b89-4cd0-ad07-485f26400cc1 req-555309b3-abc8-4823-ae32-03929d72dbef service nova] Releasing lock "refresh_cache-d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.828703] env[66641]: DEBUG oslo_concurrency.lockutils [req-d212e6be-d9e7-4810-8073-fe88e5ec88bc req-27e42045-beef-4949-baf4-0cf04d4fbd18 service nova] Releasing lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.911781] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145578, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 638.915716] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 638.915979] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 638.916211] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.236374] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145581, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.417330] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145578, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.550548} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.417593] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copied Virtual Disk [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk to [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 639.417593] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleting the datastore file [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51/tmp-sparse.vmdk {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 639.417835] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5be3cddf-e8fa-4ec3-af0c-60717a3444e7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.426929] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 639.426929] env[66641]: value = "task-5145582" [ 639.426929] env[66641]: _type = "Task" [ 639.426929] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.441024] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145582, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.733637] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145581, 'name': CreateVM_Task, 'duration_secs': 0.652467} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.733794] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 639.734786] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 639.735247] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 639.741917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.741917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.741917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 639.741917] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-087b87ee-928f-42aa-b2d4-908bae51b6d6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.748338] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 639.748338] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f9fe10-f9dd-53d7-cc62-fc5dab189a66" [ 639.748338] env[66641]: _type = "Task" [ 639.748338] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.757502] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f9fe10-f9dd-53d7-cc62-fc5dab189a66, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.871500] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Acquiring lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.871500] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.871500] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Acquiring lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.871500] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.871757] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.873139] env[66641]: INFO nova.compute.manager [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Terminating instance [ 639.938490] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145582, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.031637} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 639.939138] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 639.939138] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Moving file from [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a/cd910b37-6707-4868-b172-79fffc590a51 to [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51. {{(pid=66641) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 639.939330] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-0c3f884d-e7c7-4101-b013-c09a4d5cb09f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.952460] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 639.952460] env[66641]: value = "task-5145583" [ 639.952460] env[66641]: _type = "Task" [ 639.952460] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 639.964765] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145583, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 639.975047] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Received event network-vif-plugged-46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 639.975462] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquiring lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 639.975667] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 639.976029] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 639.976298] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] No waiting events found dispatching network-vif-plugged-46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 639.977385] env[66641]: WARNING nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Received unexpected event network-vif-plugged-46c2027b-2796-42b8-9167-719bceab5640 for instance with vm_state building and task_state spawning. [ 639.977385] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Received event network-changed-46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 639.977385] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Refreshing instance network info cache due to event network-changed-46c2027b-2796-42b8-9167-719bceab5640. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 639.977385] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquiring lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.977385] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquired lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 639.977651] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Refreshing network info cache for port 46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 640.062331] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 640.062556] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 640.259151] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 640.259151] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 640.259151] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.380165] env[66641]: DEBUG nova.compute.manager [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 640.380165] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.380165] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835b973d-e4b5-4d4f-b5f4-7b64bf927e46 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.387937] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 640.388377] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b7724d17-3a8e-48b8-aaa3-f5b714b12a2e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.395911] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Waiting for the task: (returnval){ [ 640.395911] env[66641]: value = "task-5145584" [ 640.395911] env[66641]: _type = "Task" [ 640.395911] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.407175] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Task: {'id': task-5145584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.465806] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145583, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.034496} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.465806] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] File moved {{(pid=66641) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 640.466175] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Cleaning up location [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 640.466366] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleting the datastore file [datastore1] vmware_temp/7616a257-ac5a-418d-9f66-c5f1f08c712a {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 640.466615] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d46f362e-c7c5-4d4f-bfcd-7ad2c5422c1c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.475455] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 640.475455] env[66641]: value = "task-5145585" [ 640.475455] env[66641]: _type = "Task" [ 640.475455] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 640.480389] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 640.480788] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 640.492376] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 640.565275] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 640.911040] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Task: {'id': task-5145584, 'name': PowerOffVM_Task, 'duration_secs': 0.238665} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.911315] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 640.911532] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 640.911878] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fbefed84-8e56-415b-b881-f47aa9db0d5c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.986949] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.041628} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 640.987524] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 640.988453] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b43aac7-81ab-4038-976c-81d68325e527 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.996021] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 640.996021] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 640.996021] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Deleting the datastore file [datastore2] 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 640.996021] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dcd8409d-3f2a-443b-8719-492dac0435ac {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.001149] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 641.001149] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524660dc-dd59-04e4-ccc2-28fcce238f0e" [ 641.001149] env[66641]: _type = "Task" [ 641.001149] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.011128] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Waiting for the task: (returnval){ [ 641.011128] env[66641]: value = "task-5145587" [ 641.011128] env[66641]: _type = "Task" [ 641.011128] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.018365] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524660dc-dd59-04e4-ccc2-28fcce238f0e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.024634] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Task: {'id': task-5145587, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.090487] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 641.090487] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 641.092260] env[66641]: INFO nova.compute.claims [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 641.488047] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.491233] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.529026] env[66641]: DEBUG nova.compute.manager [None req-084957f3-1f24-4f26-bd7e-2b67268233f8 tempest-ServerDiagnosticsV248Test-2136313891 tempest-ServerDiagnosticsV248Test-2136313891-project-admin] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 641.529026] env[66641]: DEBUG oslo_vmware.api [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Task: {'id': task-5145587, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.409676} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.532046] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47601133-8ab9-4117-a80c-5eb8f08cd981 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.535018] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 641.535308] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 641.535424] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 641.535597] env[66641]: INFO nova.compute.manager [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Took 1.16 seconds to destroy the instance on the hypervisor. [ 641.535855] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 641.536445] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524660dc-dd59-04e4-ccc2-28fcce238f0e, 'name': SearchDatastore_Task, 'duration_secs': 0.019734} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 641.536758] env[66641]: DEBUG nova.compute.manager [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 641.536758] env[66641]: DEBUG nova.network.neutron [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 641.537300] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 641.537606] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 641.545075] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 641.545414] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 641.546570] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 641.546935] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 641.547099] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fb8b8a67-5c60-4952-9cba-f25aef135f72 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.551258] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9dfb2daa-0948-40f4-9769-7a7b0d81d314 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.552971] env[66641]: INFO nova.compute.manager [None req-084957f3-1f24-4f26-bd7e-2b67268233f8 tempest-ServerDiagnosticsV248Test-2136313891 tempest-ServerDiagnosticsV248Test-2136313891-project-admin] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Retrieving diagnostics [ 641.554143] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da36771-bd7e-40e1-bdbe-2a47fba8498b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.558752] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 641.558752] env[66641]: value = "task-5145588" [ 641.558752] env[66641]: _type = "Task" [ 641.558752] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.595810] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 641.596024] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 641.599635] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e40d708-a789-4857-b50e-660d71763b44 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.606467] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145588, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 641.611032] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 641.611032] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233964e-2053-8a90-b4fd-7f641909c603" [ 641.611032] env[66641]: _type = "Task" [ 641.611032] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 641.622277] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233964e-2053-8a90-b4fd-7f641909c603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.071174] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145588, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.090426] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.091707] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.124673] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233964e-2053-8a90-b4fd-7f641909c603, 'name': SearchDatastore_Task, 'duration_secs': 0.020732} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.125556] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bdd3098-d08a-4d54-967e-320fb9e7acfa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.133093] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 642.133093] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52df7ffd-b366-9791-bb76-2e447f942cd2" [ 642.133093] env[66641]: _type = "Task" [ 642.133093] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.146539] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52df7ffd-b366-9791-bb76-2e447f942cd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.178154] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 642.179705] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 642.307273] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f43dc797-e2da-45af-ae99-e995fa321ebf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.314617] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe3f829-25c6-4b0c-86ce-cd6aa3f5f54a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.348901] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35618b68-2fe0-419e-b59d-c930458d1b9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.357202] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14726875-458f-473b-a9e2-fc83124d7a3d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.376708] env[66641]: DEBUG nova.compute.provider_tree [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.577446] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145588, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.667407} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.577446] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 642.577856] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 642.577856] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-baac5b09-7c2b-41c7-97c7-61d7ba53da7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.587374] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 642.587374] env[66641]: value = "task-5145589" [ 642.587374] env[66641]: _type = "Task" [ 642.587374] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.600401] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145589, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.645148] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52df7ffd-b366-9791-bb76-2e447f942cd2, 'name': SearchDatastore_Task, 'duration_secs': 0.054056} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 642.645427] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 642.645686] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 47aaed25-542b-4ceb-9adf-6a4953c8c95d/47aaed25-542b-4ceb-9adf-6a4953c8c95d.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 642.645977] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 642.647279] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 642.647708] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c7532bc-8681-4556-b206-884468e0a0b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.650916] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48139b0a-b66c-483a-ac6a-93c1b85ad635 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.661019] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 642.661019] env[66641]: value = "task-5145590" [ 642.661019] env[66641]: _type = "Task" [ 642.661019] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.663696] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 642.663937] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 642.665529] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25899bfc-0239-462a-ab93-efef493c635f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.671374] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145590, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.675550] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 642.675550] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f63621-92a2-d741-15aa-d8a2796a0be4" [ 642.675550] env[66641]: _type = "Task" [ 642.675550] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 642.686587] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f63621-92a2-d741-15aa-d8a2796a0be4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 642.713199] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.713539] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.713814] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 642.713972] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 642.714172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 642.716458] env[66641]: INFO nova.compute.manager [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Terminating instance [ 642.798345] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Updated VIF entry in instance network info cache for port 46c2027b-2796-42b8-9167-719bceab5640. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 642.800056] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Updating instance_info_cache with network_info: [{"id": "46c2027b-2796-42b8-9167-719bceab5640", "address": "fa:16:3e:d5:fb:69", "network": {"id": "7929b0b6-79ef-4cfa-a453-26990f6083fa", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-826902442-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6911bbec11fa43a4a2db344655d00d32", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b9aabc7c-0f6c-42eb-bd27-493a1496c0c8", "external-id": "nsx-vlan-transportzone-368", "segmentation_id": 368, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46c2027b-27", "ovs_interfaceid": "46c2027b-2796-42b8-9167-719bceab5640", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 642.883025] env[66641]: DEBUG nova.scheduler.client.report [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 642.890433] env[66641]: DEBUG nova.network.neutron [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 643.102283] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145589, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069117} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.103116] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 643.103882] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d6d346-b0fd-48ca-a2e1-0d2eac95d0b3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.133124] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 643.133124] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21981e04-f08c-4f20-91a6-a679da5343c6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.156868] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 643.156868] env[66641]: value = "task-5145591" [ 643.156868] env[66641]: _type = "Task" [ 643.156868] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.174421] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145590, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.174562] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145591, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.189798] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f63621-92a2-d741-15aa-d8a2796a0be4, 'name': SearchDatastore_Task, 'duration_secs': 0.020153} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.191050] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-032119a9-a430-4db6-86a4-4e71a14f9026 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.198128] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "6b19204d-032c-410b-bb84-1af23d142edc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.198670] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.202638] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 643.202638] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5220874c-6806-03b5-085c-a93124d94a1f" [ 643.202638] env[66641]: _type = "Task" [ 643.202638] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.212895] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5220874c-6806-03b5-085c-a93124d94a1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.219675] env[66641]: DEBUG nova.compute.manager [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 643.219999] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 643.221009] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-669b1e3f-4a72-4828-8e37-88e5f9c5a6be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.231329] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 643.231329] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bc020ad-ce20-422a-8fc7-bf9fe784bc7b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.240149] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 643.240149] env[66641]: value = "task-5145592" [ 643.240149] env[66641]: _type = "Task" [ 643.240149] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.255373] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.306019] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Releasing lock "refresh_cache-47aaed25-542b-4ceb-9adf-6a4953c8c95d" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.306019] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Received event network-vif-plugged-3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 643.306019] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquiring lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.306019] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.306019] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.306324] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] No waiting events found dispatching network-vif-plugged-3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 643.306324] env[66641]: WARNING nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Received unexpected event network-vif-plugged-3e466382-fc5f-4ac9-90dc-05e1d9f93c48 for instance with vm_state building and task_state deleting. [ 643.306324] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Received event network-changed-3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 643.306324] env[66641]: DEBUG nova.compute.manager [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Refreshing instance network info cache due to event network-changed-3e466382-fc5f-4ac9-90dc-05e1d9f93c48. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 643.306324] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquiring lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.306813] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Acquired lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 643.308603] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Refreshing network info cache for port 3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 643.387274] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.297s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 643.387807] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 643.394570] env[66641]: INFO nova.compute.manager [-] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Took 1.86 seconds to deallocate network for instance. [ 643.677470] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145590, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.803181} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.677808] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145591, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.678096] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 47aaed25-542b-4ceb-9adf-6a4953c8c95d/47aaed25-542b-4ceb-9adf-6a4953c8c95d.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 643.678316] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 643.678552] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c81f2198-d77a-48f6-965c-1389fc4fcb9e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.690053] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 643.690053] env[66641]: value = "task-5145593" [ 643.690053] env[66641]: _type = "Task" [ 643.690053] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.705053] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.709279] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 643.719959] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5220874c-6806-03b5-085c-a93124d94a1f, 'name': SearchDatastore_Task, 'duration_secs': 0.058033} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.720326] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 643.720872] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc/d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 643.722433] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3df17237-3a49-46fa-abde-eaf51157fdbf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.735692] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 643.735692] env[66641]: value = "task-5145594" [ 643.735692] env[66641]: _type = "Task" [ 643.735692] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.748343] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145594, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.756295] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145592, 'name': PowerOffVM_Task, 'duration_secs': 0.369399} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 643.756634] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 643.756899] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 643.757189] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4c869aa-96a2-47c2-afe7-12a6e0c50802 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.812318] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.814337] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.828670] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.829377] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.832296] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 643.832603] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 643.833097] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Deleting the datastore file [datastore2] 25203771-cf4d-4fc1-a1bf-f2a0200296b7 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 643.833614] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ec8be4b1-0fbc-4a3e-b62a-97c48d616449 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.842448] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for the task: (returnval){ [ 643.842448] env[66641]: value = "task-5145596" [ 643.842448] env[66641]: _type = "Task" [ 643.842448] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 643.853509] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 643.893504] env[66641]: DEBUG nova.compute.utils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 643.895355] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 643.895642] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 643.896118] env[66641]: WARNING neutronclient.v2_0.client [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.896629] env[66641]: WARNING neutronclient.v2_0.client [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 643.897481] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 643.898015] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 643.910982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 643.911599] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 643.911998] env[66641]: DEBUG nova.objects.instance [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lazy-loading 'resources' on Instance uuid 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 644.170446] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145591, 'name': ReconfigVM_Task, 'duration_secs': 0.848797} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.170803] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfigured VM instance instance-00000006 to attach disk [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 644.171883] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e29ceb2-5258-4f5f-84f8-8aece7f432a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.180780] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 644.180780] env[66641]: value = "task-5145597" [ 644.180780] env[66641]: _type = "Task" [ 644.180780] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.192406] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145597, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.202629] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145593, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071509} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.202629] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 644.203855] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140edb9c-7629-4ed5-87f5-2608f40a7c8c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.238209] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Reconfiguring VM instance instance-00000007 to attach disk [datastore1] 47aaed25-542b-4ceb-9adf-6a4953c8c95d/47aaed25-542b-4ceb-9adf-6a4953c8c95d.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 644.238209] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.238209] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.239474] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.239474] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 644.239474] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 644.240560] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32a92ec8-3e4e-409c-a353-9c64404b9000 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.257845] env[66641]: INFO nova.compute.manager [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Terminating instance [ 644.273199] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145594, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.275081] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 644.275081] env[66641]: value = "task-5145598" [ 644.275081] env[66641]: _type = "Task" [ 644.275081] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.283166] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 644.334948] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 644.356255] env[66641]: DEBUG oslo_vmware.api [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Task: {'id': task-5145596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323684} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.359191] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 644.359191] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 644.359191] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.359191] env[66641]: INFO nova.compute.manager [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Took 1.14 seconds to destroy the instance on the hypervisor. [ 644.359191] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 644.359515] env[66641]: DEBUG nova.compute.manager [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 644.359515] env[66641]: DEBUG nova.network.neutron [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 644.359856] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.360806] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.408717] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 644.656056] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-745466da-43e3-4b92-ae49-8fbad2a8716e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.666131] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6600ee-ff01-4ddc-b9df-f4b672c8392d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.676655] env[66641]: DEBUG nova.policy [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd1dfe5b347f4871b7d6c0fb6770e1b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8449c95dd0f74492929e4d1d492ac9d2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 644.708167] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 644.708451] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 644.720048] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16c61d7e-8dfc-4834-9286-60953476cfd8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.731514] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145597, 'name': Rename_Task, 'duration_secs': 0.18351} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.734055] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 644.736202] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-060e3793-b523-4e6b-99b6-bc29ba5717d6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.739387] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d5dbd5-88e0-44ce-ba5e-3a65c1f3e198 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.755621] env[66641]: DEBUG nova.compute.provider_tree [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.759372] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 644.759372] env[66641]: value = "task-5145599" [ 644.759372] env[66641]: _type = "Task" [ 644.759372] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.768514] env[66641]: DEBUG nova.compute.manager [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 644.768733] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.769853] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd2b0f9-1e85-4d68-94be-b88621ffe0c7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.777679] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145594, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.615143} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.784996] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc/d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 644.785283] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 644.785465] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145599, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.788536] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1609ed53-8fd7-46d7-b02d-20c9bae0eb60 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.791282] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 644.792259] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e2e851bb-21e5-4eec-9f98-8741fceeb4b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.800126] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145598, 'name': ReconfigVM_Task, 'duration_secs': 0.380623} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 644.801932] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Reconfigured VM instance instance-00000007 to attach disk [datastore1] 47aaed25-542b-4ceb-9adf-6a4953c8c95d/47aaed25-542b-4ceb-9adf-6a4953c8c95d.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 644.802832] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 644.802832] env[66641]: value = "task-5145600" [ 644.802832] env[66641]: _type = "Task" [ 644.802832] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.804419] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ae27250-cb81-4521-bf5e-8372508c2196 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.806704] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 644.806704] env[66641]: value = "task-5145601" [ 644.806704] env[66641]: _type = "Task" [ 644.806704] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.817344] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 644.817344] env[66641]: value = "task-5145602" [ 644.817344] env[66641]: _type = "Task" [ 644.817344] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 644.828727] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145601, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.829172] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145600, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.836052] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145602, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 644.858941] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 645.259367] env[66641]: DEBUG nova.scheduler.client.report [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 645.275272] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145599, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.320550] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145601, 'name': PowerOffVM_Task, 'duration_secs': 0.235845} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.326882] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 645.327178] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 645.327580] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145600, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081684} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.327728] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab6fac82-24b9-4ea1-9990-e21d62d83234 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.329327] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 645.330515] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cc3b87-ddd2-4704-84dd-4bc7efcea458 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.339344] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145602, 'name': Rename_Task, 'duration_secs': 0.173366} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.339344] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 645.339344] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-069d48a0-2ea7-49f2-aba9-7674179cba24 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.361512] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Reconfiguring VM instance instance-00000008 to attach disk [datastore1] d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc/d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 645.362386] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7418ff7-9e5c-44a1-9900-42e9cc766988 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.379396] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 645.379396] env[66641]: value = "task-5145604" [ 645.379396] env[66641]: _type = "Task" [ 645.379396] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.386616] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 645.386616] env[66641]: value = "task-5145605" [ 645.386616] env[66641]: _type = "Task" [ 645.386616] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.390264] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145604, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.399947] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145605, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.421332] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 645.430562] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 645.430562] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 645.430789] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Deleting the datastore file [datastore2] ccf46c5c-1c79-4672-ad9f-ea61042097d5 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 645.431136] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4681db92-716c-4030-ab6d-8cb583e269e9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.442784] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for the task: (returnval){ [ 645.442784] env[66641]: value = "task-5145606" [ 645.442784] env[66641]: _type = "Task" [ 645.442784] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.453914] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145606, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.464663] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 645.464976] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 645.465167] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 645.465348] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 645.465562] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 645.465702] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 645.465979] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.466239] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 645.467055] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 645.467055] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 645.467055] env[66641]: DEBUG nova.virt.hardware [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 645.467849] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2247fc70-1b1c-47e9-aca3-0c6f79981649 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.478636] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-761deb00-790c-4829-a204-2009e37aee41 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.767201] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.855s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 645.774306] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.491s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 645.777105] env[66641]: INFO nova.compute.claims [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.790571] env[66641]: DEBUG oslo_vmware.api [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145599, 'name': PowerOnVM_Task, 'duration_secs': 0.589576} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.790912] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 645.791127] env[66641]: INFO nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Took 17.37 seconds to spawn the instance on the hypervisor. [ 645.791309] env[66641]: DEBUG nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 645.792153] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2537cf5-04dd-46e0-871a-c596c01ef291 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.816347] env[66641]: INFO nova.scheduler.client.report [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Deleted allocations for instance 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c [ 645.899230] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145604, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.906444] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145605, 'name': ReconfigVM_Task, 'duration_secs': 0.437227} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.906732] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Reconfigured VM instance instance-00000008 to attach disk [datastore1] d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc/d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 645.907447] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39bf91fa-6883-4452-a294-9f3ab301d5de {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.916020] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 645.916020] env[66641]: value = "task-5145607" [ 645.916020] env[66641]: _type = "Task" [ 645.916020] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 645.926931] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145607, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 645.954678] env[66641]: DEBUG oslo_vmware.api [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Task: {'id': task-5145606, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184169} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 645.955039] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 645.955039] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 645.955275] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.955446] env[66641]: INFO nova.compute.manager [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Took 1.19 seconds to destroy the instance on the hypervisor. [ 645.955722] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 645.955935] env[66641]: DEBUG nova.compute.manager [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 645.956040] env[66641]: DEBUG nova.network.neutron [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 645.956804] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 645.956907] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.008241] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.009923] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.070855] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Successfully created port: aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 646.107377] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.109031] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.316246] env[66641]: INFO nova.compute.manager [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Took 28.85 seconds to build instance. [ 646.328949] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c7058641-b2d1-4d5d-a90d-8ae11ee978f1 tempest-DeleteServersAdminTestJSON-1534534325 tempest-DeleteServersAdminTestJSON-1534534325-project-admin] Lock "420544a8-1a02-42d5-8a9a-e3e8b6a11a0c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.459s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.394927] env[66641]: DEBUG oslo_vmware.api [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145604, 'name': PowerOnVM_Task, 'duration_secs': 0.583956} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.395615] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 646.395615] env[66641]: INFO nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Took 15.68 seconds to spawn the instance on the hypervisor. [ 646.395615] env[66641]: DEBUG nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 646.396684] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dac98e7a-42d8-4d3f-93c6-3d6247ae2dfb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.432077] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145607, 'name': Rename_Task, 'duration_secs': 0.335645} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 646.432077] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 646.432077] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3efc486-e9b8-4eeb-98eb-a87e77e1dbc1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.435268] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 646.435613] env[66641]: WARNING openstack [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 646.455085] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 646.455085] env[66641]: value = "task-5145608" [ 646.455085] env[66641]: _type = "Task" [ 646.455085] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 646.466855] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145608, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.480076] env[66641]: DEBUG nova.network.neutron [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.654525] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updated VIF entry in instance network info cache for port 3e466382-fc5f-4ac9-90dc-05e1d9f93c48. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 646.654848] env[66641]: DEBUG nova.network.neutron [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updating instance_info_cache with network_info: [{"id": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "address": "fa:16:3e:a2:b8:43", "network": {"id": "b6204f09-6035-4360-9da5-498dda901ff2", "bridge": "br-int", "label": "tempest-ServersTestJSON-324682247-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e25db00afd9044e3940b419d37751237", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "47ca1ce6-8148-48d5-bcfe-89e39b73914e", "external-id": "nsx-vlan-transportzone-259", "segmentation_id": 259, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e466382-fc", "ovs_interfaceid": "3e466382-fc5f-4ac9-90dc-05e1d9f93c48", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.735933] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 646.736274] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 646.821106] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cd1b09d-dd58-47c5-8615-1922ec745c71 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.380s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 646.885824] env[66641]: DEBUG nova.network.neutron [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 646.923075] env[66641]: INFO nova.compute.manager [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Took 25.43 seconds to build instance. [ 646.968935] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145608, 'name': PowerOnVM_Task} progress is 79%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 646.983882] env[66641]: INFO nova.compute.manager [-] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Took 2.62 seconds to deallocate network for instance. [ 647.077718] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3502868-52fe-4907-af64-7448da3c3fe3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.088289] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53406b5-108d-4bb2-a864-1d71568b0bd0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.136147] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79cbb819-0312-48dc-aa56-695286cc0d5e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.145911] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-891b0cbc-f48c-4ce6-aae5-504d2f16bba6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.163474] env[66641]: DEBUG oslo_concurrency.lockutils [req-f11b3b95-0f9e-49bd-99ac-af8c94e0ba45 req-7cfca618-a8c9-4bfd-a5cc-5cfb73dc5643 service nova] Releasing lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 647.164269] env[66641]: DEBUG nova.compute.provider_tree [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.240759] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 647.387689] env[66641]: INFO nova.compute.manager [-] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Took 1.43 seconds to deallocate network for instance. [ 647.424536] env[66641]: DEBUG oslo_concurrency.lockutils [None req-05a94c65-e5dc-46f4-a9d9-fe20a0cca113 tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.542s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.472570] env[66641]: DEBUG oslo_vmware.api [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145608, 'name': PowerOnVM_Task, 'duration_secs': 0.944084} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 647.472570] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 647.472934] env[66641]: INFO nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Took 13.93 seconds to spawn the instance on the hypervisor. [ 647.473042] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 647.474379] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4e0b76-6815-4695-b8bc-096d734573a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.499866] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.670027] env[66641]: DEBUG nova.scheduler.client.report [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 647.755764] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Successfully updated port: aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 647.776539] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.842252] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.842608] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.844097] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.844097] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 647.844097] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 647.845802] env[66641]: INFO nova.compute.manager [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Terminating instance [ 647.895286] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 647.991068] env[66641]: DEBUG nova.compute.utils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Conflict updating instance d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} {{(pid=66641) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 647.993775] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance disappeared during build. {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2556}} [ 647.993973] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Unplugging VIFs for instance {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3050}} [ 647.995020] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3073}} [ 647.995020] env[66641]: DEBUG nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 647.995020] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 647.997738] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 647.998266] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.175821] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.402s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 648.176484] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 648.180035] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.320s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 648.181749] env[66641]: INFO nova.compute.claims [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.240038] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.244643] env[66641]: WARNING openstack [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.266719] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.266719] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.266719] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 648.316038] env[66641]: DEBUG nova.compute.manager [req-66668a52-35e6-4993-92ba-8af9d84003cb req-9d143561-af6a-4d3a-a3c4-e9968eec60e8 service nova] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Received event network-vif-deleted-88f6140a-2061-43ef-9eed-eea8ab8878c5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 648.349970] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "refresh_cache-4ebc0f2b-bb79-4941-b8f2-082560c80f8b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.350140] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquired lock "refresh_cache-4ebc0f2b-bb79-4941-b8f2-082560c80f8b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 648.350357] env[66641]: DEBUG nova.network.neutron [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 648.686601] env[66641]: DEBUG nova.compute.utils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 648.690830] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 648.690830] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 648.691152] env[66641]: WARNING neutronclient.v2_0.client [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.691475] env[66641]: WARNING neutronclient.v2_0.client [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 648.692611] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.693218] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.766686] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.767350] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.772782] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 648.854311] env[66641]: WARNING openstack [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 648.854775] env[66641]: WARNING openstack [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 648.860091] env[66641]: DEBUG nova.network.neutron [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 649.164600] env[66641]: DEBUG nova.policy [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bc117278b9d741a199c56536da2d2504', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '123276c617f54856923563091fd1842c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 649.191719] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 649.257309] env[66641]: DEBUG nova.network.neutron [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 649.443687] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-516d7f83-b6d6-4cab-901f-633db155d9bc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.453205] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e32958ca-ff28-441f-b33d-bd0faea81594 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.501441] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a792a41f-ae9f-4bba-b302-c2d5fce04f70 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.513357] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72416516-e334-4828-baff-59eb5611e936 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.533071] env[66641]: DEBUG nova.compute.provider_tree [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.765216] env[66641]: INFO nova.compute.manager [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Took 1.77 seconds to deallocate network for instance. [ 649.889910] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Successfully created port: 9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 649.903862] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 649.904286] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.042288] env[66641]: DEBUG nova.scheduler.client.report [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 650.203906] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 650.241715] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 650.241961] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 650.242134] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 650.242316] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 650.242456] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 650.242591] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 650.242829] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.242987] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 650.243569] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 650.243925] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 650.243925] env[66641]: DEBUG nova.virt.hardware [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 650.244878] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eaad698-1dc0-4b16-824c-b6eca0381d5c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.255327] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342324cc-c611-422b-b4ce-f9c7b5e7cddf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.429344] env[66641]: DEBUG nova.network.neutron [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.551430] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.551951] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 650.554420] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.055s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.554619] env[66641]: DEBUG nova.objects.instance [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lazy-loading 'resources' on Instance uuid 25203771-cf4d-4fc1-a1bf-f2a0200296b7 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 650.613929] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 650.614355] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 650.750392] env[66641]: DEBUG nova.network.neutron [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Updating instance_info_cache with network_info: [{"id": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "address": "fa:16:3e:d8:c3:ea", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa8a5251-c9", "ovs_interfaceid": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 650.808577] env[66641]: INFO nova.scheduler.client.report [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Deleted allocations for instance d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc [ 650.809172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e48708ba-1486-47bf-b7c3-8d7a36c1fe8c tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.619s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.809555] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 17.116s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.809555] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 650.809963] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 650.810521] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 650.812692] env[66641]: INFO nova.compute.manager [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Terminating instance [ 650.933244] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Releasing lock "refresh_cache-4ebc0f2b-bb79-4941-b8f2-082560c80f8b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 650.935286] env[66641]: DEBUG nova.compute.manager [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 650.935763] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 650.937333] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96d591a2-fc77-43d6-8812-4c69a2c90f4b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.948038] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 650.948389] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-03f40360-e278-4c9c-b4e6-c15fe3eb74e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.957376] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 650.957376] env[66641]: value = "task-5145610" [ 650.957376] env[66641]: _type = "Task" [ 650.957376] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 650.971536] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145610, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.021843] env[66641]: DEBUG nova.compute.manager [req-02a231bc-460e-41af-829b-882268900cac req-d08612ed-3eaf-4af4-bd25-89fb4ff15ae7 service nova] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Received event network-vif-deleted-0c57e5de-8952-4f18-8e92-c4cc1fb19ef7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 651.056944] env[66641]: DEBUG nova.compute.utils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 651.063021] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 651.254548] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 651.254548] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance network_info: |[{"id": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "address": "fa:16:3e:d8:c3:ea", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa8a5251-c9", "ovs_interfaceid": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 651.254978] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:c3:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 651.266737] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Creating folder: Project (8449c95dd0f74492929e4d1d492ac9d2). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 651.274277] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0cf4a16-3f08-4b22-921d-ee2ad5a14265 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.287856] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Created folder: Project (8449c95dd0f74492929e4d1d492ac9d2) in parent group-v1000566. [ 651.288088] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Creating folder: Instances. Parent ref: group-v1000594. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 651.288357] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f3686058-d96c-4055-aa71-df45e4aebcb0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.296482] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c02fb6-7791-434c-a8f5-6e557188809a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.305382] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Created folder: Instances in parent group-v1000594. [ 651.305382] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 651.305382] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 651.305382] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15865c2f-a54e-42be-a17d-56824dbabc39 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.323215] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.323435] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquired lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.323607] env[66641]: DEBUG nova.network.neutron [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 651.328397] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc9f106-6776-47c8-8830-19df14fbf486 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.334287] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 651.334287] env[66641]: value = "task-5145616" [ 651.334287] env[66641]: _type = "Task" [ 651.334287] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.375543] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f50598ef-fca2-40e3-92fa-ede90a5a77c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.382065] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145616, 'name': CreateVM_Task} progress is 15%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.389876] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-089b4b96-e25b-4883-ac55-1b0ceb72b3b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.407874] env[66641]: DEBUG nova.compute.provider_tree [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.473109] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "7fa45876-59c2-4fa2-978b-46067931e091" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 651.474371] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "7fa45876-59c2-4fa2-978b-46067931e091" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 651.483551] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145610, 'name': PowerOffVM_Task, 'duration_secs': 0.190004} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.484337] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 651.484337] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 651.484337] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ac25b34-7f5c-4302-b0cd-7fa60cd510b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.516011] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 651.516618] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 651.516878] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleting the datastore file [datastore2] 4ebc0f2b-bb79-4941-b8f2-082560c80f8b {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 651.518855] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6657991b-d0b5-4240-8f5d-085e10dba2d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.531759] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for the task: (returnval){ [ 651.531759] env[66641]: value = "task-5145618" [ 651.531759] env[66641]: _type = "Task" [ 651.531759] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.551249] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145618, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.564628] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 651.833247] env[66641]: WARNING openstack [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.833833] env[66641]: WARNING openstack [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.844516] env[66641]: DEBUG nova.network.neutron [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 651.860394] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145616, 'name': CreateVM_Task, 'duration_secs': 0.505835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 651.861997] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 651.861997] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 651.862312] env[66641]: WARNING openstack [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 651.870141] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.872253] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 651.872253] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 651.872253] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-15748206-fb4c-466d-bfe1-2507d444302e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.878258] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 651.878258] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52406880-086d-c4ca-3824-78790522eb5d" [ 651.878258] env[66641]: _type = "Task" [ 651.878258] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 651.890082] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52406880-086d-c4ca-3824-78790522eb5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 651.914972] env[66641]: DEBUG nova.scheduler.client.report [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 651.961598] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Successfully updated port: 9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 651.976669] env[66641]: DEBUG nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 652.047978] env[66641]: DEBUG oslo_vmware.api [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Task: {'id': task-5145618, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19105} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.048713] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 652.049011] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 652.049199] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 652.049381] env[66641]: INFO nova.compute.manager [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 652.049722] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 652.051036] env[66641]: DEBUG nova.compute.manager [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 652.051158] env[66641]: DEBUG nova.network.neutron [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 652.051766] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.051899] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.070020] env[66641]: DEBUG nova.network.neutron [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.079219] env[66641]: DEBUG nova.network.neutron [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 652.081138] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.081138] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.297478] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.297856] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.395518] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52406880-086d-c4ca-3824-78790522eb5d, 'name': SearchDatastore_Task, 'duration_secs': 0.03794} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.396725] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.397133] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 652.397816] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.398143] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.398436] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 652.398791] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-86a9c79e-43a7-41b6-a979-363b18e95ffc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.409038] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 652.409179] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 652.410043] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92ba4db9-13c8-4ff3-a3e7-451f6fd99749 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.416815] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 652.416815] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e395c6-a7c5-89f5-ee6b-b8a2f8536970" [ 652.416815] env[66641]: _type = "Task" [ 652.416815] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.434131] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.879s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 652.437383] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e395c6-a7c5-89f5-ee6b-b8a2f8536970, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.437383] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.661s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 652.440086] env[66641]: INFO nova.compute.claims [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.465087] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.465087] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquired lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 652.465087] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 652.493667] env[66641]: INFO nova.scheduler.client.report [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Deleted allocations for instance 25203771-cf4d-4fc1-a1bf-f2a0200296b7 [ 652.522783] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 652.578346] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Releasing lock "refresh_cache-d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 652.579316] env[66641]: DEBUG nova.compute.manager [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 652.579316] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.580442] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 652.585177] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319842ab-42da-4de7-82e5-3da04ac703dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.589402] env[66641]: DEBUG nova.network.neutron [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 652.598404] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 652.598722] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c6feb20-6941-40b8-b303-b4b4f4ebdacc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.608026] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 652.608026] env[66641]: value = "task-5145619" [ 652.608026] env[66641]: _type = "Task" [ 652.608026] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.623160] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145619, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.628668] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 652.628979] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 652.629134] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 652.629237] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 652.629395] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 652.629696] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 652.630998] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.630998] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 652.630998] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 652.630998] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 652.630998] env[66641]: DEBUG nova.virt.hardware [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 652.632834] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d13e05-d3ef-49cf-ae10-51befcb53db1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.643280] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-750200c8-0055-4f23-ab3c-38748b307195 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.662435] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 652.672914] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Creating folder: Project (7f8e91d542844e1fad6ba8c274d0756d). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 652.673369] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-045401cc-657b-449d-9e63-dc447cfd0977 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.688551] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Created folder: Project (7f8e91d542844e1fad6ba8c274d0756d) in parent group-v1000566. [ 652.688753] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Creating folder: Instances. Parent ref: group-v1000597. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 652.689039] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e7f19748-c6c0-4e64-8559-bc8be5a03465 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.702711] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Created folder: Instances in parent group-v1000597. [ 652.704448] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 652.704448] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 652.704448] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-862e87fa-5673-43ab-a692-05c32460dbcf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.729201] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 652.729201] env[66641]: value = "task-5145622" [ 652.729201] env[66641]: _type = "Task" [ 652.729201] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.739197] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145622, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.801227] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 652.930588] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e395c6-a7c5-89f5-ee6b-b8a2f8536970, 'name': SearchDatastore_Task, 'duration_secs': 0.025958} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 652.932653] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ae217f3-f625-42f6-8b8c-8e2b5049f0b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.938920] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 652.938920] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233c874-8cdc-f252-f48a-145921852a2a" [ 652.938920] env[66641]: _type = "Task" [ 652.938920] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 652.954413] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233c874-8cdc-f252-f48a-145921852a2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 652.968387] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 652.972019] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 652.977038] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 653.004338] env[66641]: DEBUG nova.compute.manager [req-cb545194-4342-4ff2-a19d-d3aefb71374e req-ccf4c738-c56f-4904-8775-4c6c1ee4cc5a service nova] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Received event network-vif-deleted-361c7a50-fe6b-4975-bf4e-5bceab411bc7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 653.004338] env[66641]: DEBUG oslo_concurrency.lockutils [None req-79550bb8-817c-4de6-8591-f47680336f74 tempest-ServerDiagnosticsNegativeTest-907938446 tempest-ServerDiagnosticsNegativeTest-907938446-project-member] Lock "25203771-cf4d-4fc1-a1bf-f2a0200296b7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.291s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 653.068939] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.069448] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.095146] env[66641]: INFO nova.compute.manager [-] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Took 1.04 seconds to deallocate network for instance. [ 653.126677] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145619, 'name': PowerOffVM_Task, 'duration_secs': 0.339793} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.126677] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 653.126677] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 653.132316] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fcdaa88-bece-4183-afc1-3d729d4b35a9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.169342] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.169871] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.220265] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 653.220792] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 653.221173] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Deleting the datastore file [datastore1] d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 653.221916] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9359ba6-eb8d-43c8-a44b-cb14d729706b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.234283] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for the task: (returnval){ [ 653.234283] env[66641]: value = "task-5145624" [ 653.234283] env[66641]: _type = "Task" [ 653.234283] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.242984] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145622, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.248341] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145624, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.334533] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.337018] env[66641]: DEBUG nova.network.neutron [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Updating instance_info_cache with network_info: [{"id": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "address": "fa:16:3e:bf:1b:08", "network": {"id": "0a7d6c39-f10c-4cd3-9740-fc9c04d9962f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-623599665-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "123276c617f54856923563091fd1842c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c5aba0e-5f", "ovs_interfaceid": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 653.463753] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233c874-8cdc-f252-f48a-145921852a2a, 'name': SearchDatastore_Task, 'duration_secs': 0.028095} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.464224] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.464613] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 653.465494] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-09a5a2cf-d981-4f51-94e8-f902621c2744 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.474834] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 653.474834] env[66641]: value = "task-5145625" [ 653.474834] env[66641]: _type = "Task" [ 653.474834] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.491738] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.610296] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 653.719821] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394bd4e9-1609-495c-9b13-e5273d66761b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.732638] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0443cdf-2936-442c-b8f4-279f273bc3e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.743649] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145622, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.775435] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ce5094-0df5-44dc-89f6-48a0a157cf53 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.781282] env[66641]: DEBUG oslo_vmware.api [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Task: {'id': task-5145624, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.222822} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 653.781979] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 653.782183] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 653.782355] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.782609] env[66641]: INFO nova.compute.manager [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Took 1.20 seconds to destroy the instance on the hypervisor. [ 653.782865] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 653.783124] env[66641]: DEBUG nova.compute.manager [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 653.783229] env[66641]: DEBUG nova.network.neutron [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 653.783768] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.784610] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.794546] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f7c630-89be-467b-81ba-858fb33ef654 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.811798] env[66641]: DEBUG nova.compute.provider_tree [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.827081] env[66641]: DEBUG nova.network.neutron [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 653.827684] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 653.827946] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 653.843128] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Releasing lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 653.843479] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Instance network_info: |[{"id": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "address": "fa:16:3e:bf:1b:08", "network": {"id": "0a7d6c39-f10c-4cd3-9740-fc9c04d9962f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-623599665-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "123276c617f54856923563091fd1842c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c5aba0e-5f", "ovs_interfaceid": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 653.843944] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:1b:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '29945b49-d2b8-449d-9531-437917f49839', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 653.853057] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Creating folder: Project (123276c617f54856923563091fd1842c). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 653.853920] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04151a7e-ee1d-474b-a267-981ebe05dd8c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.866563] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Created folder: Project (123276c617f54856923563091fd1842c) in parent group-v1000566. [ 653.866771] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Creating folder: Instances. Parent ref: group-v1000600. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 653.867621] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-596f1722-b7ac-409a-912d-5fa43d08f428 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.880714] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Created folder: Instances in parent group-v1000600. [ 653.881377] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 653.881677] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 653.881940] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0564f08f-7991-4cf4-9bfd-bc77a3db9c1c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.906174] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 653.906174] env[66641]: value = "task-5145628" [ 653.906174] env[66641]: _type = "Task" [ 653.906174] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 653.919288] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145628, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 653.990275] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145625, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.243302] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145622, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.317034] env[66641]: DEBUG nova.scheduler.client.report [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 654.335180] env[66641]: DEBUG nova.network.neutron [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 654.421642] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145628, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.492228] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145625, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.750128] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145622, 'name': CreateVM_Task, 'duration_secs': 1.785408} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.750401] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 654.750982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.751783] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 654.752567] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 654.753364] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11ff64b2-7acc-40ec-a511-524eaaf4090c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.762399] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 654.762399] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ccb052-1240-6e6d-61c7-764378573448" [ 654.762399] env[66641]: _type = "Task" [ 654.762399] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 654.775843] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ccb052-1240-6e6d-61c7-764378573448, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 654.788661] env[66641]: DEBUG nova.compute.manager [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Stashing vm_state: active {{(pid=66641) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 654.822228] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.384s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 654.822430] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 654.827353] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.932s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.830057] env[66641]: DEBUG nova.objects.instance [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lazy-loading 'resources' on Instance uuid ccf46c5c-1c79-4672-ad9f-ea61042097d5 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 654.839030] env[66641]: INFO nova.compute.manager [-] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Took 1.06 seconds to deallocate network for instance. [ 654.908080] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.908336] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.924513] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145628, 'name': CreateVM_Task, 'duration_secs': 0.569006} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.924700] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 654.927904] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 654.928330] env[66641]: WARNING openstack [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 654.937378] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.957846] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 654.958083] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 654.988899] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145625, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.379133} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 654.989197] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 654.989685] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 654.990200] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-756d3ef3-0b01-4411-9364-ef561c362768 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.998918] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 654.998918] env[66641]: value = "task-5145630" [ 654.998918] env[66641]: _type = "Task" [ 654.998918] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.009816] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.010450] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.016570] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145630, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.205987] env[66641]: DEBUG nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Received event network-vif-plugged-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 655.205987] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Acquiring lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.206536] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.208972] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.002s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 655.209198] env[66641]: DEBUG nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] No waiting events found dispatching network-vif-plugged-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 655.212337] env[66641]: WARNING nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Received unexpected event network-vif-plugged-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 for instance with vm_state building and task_state spawning. [ 655.212337] env[66641]: DEBUG nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Received event network-changed-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 655.212337] env[66641]: DEBUG nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Refreshing instance network info cache due to event network-changed-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 655.212337] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Acquiring lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.212337] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Acquired lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.212510] env[66641]: DEBUG nova.network.neutron [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Refreshing network info cache for port aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 655.221758] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "22899269-28fc-446e-9533-1f6862a9e8db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.223564] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 655.275512] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ccb052-1240-6e6d-61c7-764378573448, 'name': SearchDatastore_Task, 'duration_secs': 0.065169} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.275512] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.275512] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 655.275512] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.276986] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.276986] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 655.276986] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.276986] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 655.276986] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f526b30a-e8fb-477f-99c1-f96c27d3b70d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.277471] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa971ec8-a641-4910-b974-320433342bd1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.283609] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 655.283609] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5211b940-5a21-34d2-df20-5e35853be0a6" [ 655.283609] env[66641]: _type = "Task" [ 655.283609] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.292629] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5211b940-5a21-34d2-df20-5e35853be0a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.315097] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.332022] env[66641]: DEBUG nova.compute.utils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 655.336515] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 655.336846] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 655.337353] env[66641]: WARNING neutronclient.v2_0.client [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.337839] env[66641]: WARNING neutronclient.v2_0.client [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 655.338959] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.339126] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.350146] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.414371] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 655.425660] env[66641]: DEBUG nova.policy [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5c60265af66d43d18b4b48ec1f0e2d05', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '91b733f620ed4947975d30e02ea73e24', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 655.441811] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 655.441897] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 655.442972] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05b6cded-b51c-486b-8de0-7f51d96b4b45 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.454088] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 655.454088] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5253cfdc-4c98-1239-174e-6637802679db" [ 655.454088] env[66641]: _type = "Task" [ 655.454088] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.466758] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 655.471299] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5253cfdc-4c98-1239-174e-6637802679db, 'name': SearchDatastore_Task, 'duration_secs': 0.011434} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.476358] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f1f2f723-720c-4f81-bfaa-10f72f3c6d04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.482541] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 655.482541] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f8d887-5db6-dcf8-a6c4-e7feaf205d3b" [ 655.482541] env[66641]: _type = "Task" [ 655.482541] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.493764] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f8d887-5db6-dcf8-a6c4-e7feaf205d3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.511193] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145630, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091203} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.513053] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 655.514210] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31aaed1f-6153-4111-b23d-c92220a8eef0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.518361] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 655.542354] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Reconfiguring VM instance instance-00000009 to attach disk [datastore2] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 655.546558] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e6b8d564-56a6-490b-a0f2-170a050cec87 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.572086] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 655.572086] env[66641]: value = "task-5145631" [ 655.572086] env[66641]: _type = "Task" [ 655.572086] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 655.583503] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145631, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 655.675969] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f292cd5-8f4b-4028-8825-180c6df85850 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.683908] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c65b37-7542-4f4c-9812-32906d3ad4e5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.716274] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12fb0c3a-61a0-4ca7-9648-b8d5a12de7ba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.726412] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 655.727443] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 655.733669] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 655.741963] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4201673-3742-421a-9071-8f9e4239a4e2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.757883] env[66641]: DEBUG nova.compute.provider_tree [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.797396] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5211b940-5a21-34d2-df20-5e35853be0a6, 'name': SearchDatastore_Task, 'duration_secs': 0.167828} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.797734] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.797939] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 655.798160] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.835350] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 655.877037] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Successfully created port: d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 655.938510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.989588] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 655.997219] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f8d887-5db6-dcf8-a6c4-e7feaf205d3b, 'name': SearchDatastore_Task, 'duration_secs': 0.011743} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 655.997602] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 655.998577] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 655.998970] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 655.999217] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 655.999432] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ffa4545b-13fd-4dd3-ae12-b77e91b337e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.001720] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f044450-5a74-437b-a1a8-97647064f694 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.010473] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 656.010473] env[66641]: value = "task-5145632" [ 656.010473] env[66641]: _type = "Task" [ 656.010473] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.016664] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 656.017143] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 656.018606] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c34135d1-595b-4d34-b77d-6126fbfbc961 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.026502] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145632, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.030796] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 656.030796] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217d37c-8ec7-a7ad-a445-c119c8c2ddde" [ 656.030796] env[66641]: _type = "Task" [ 656.030796] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.042066] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217d37c-8ec7-a7ad-a445-c119c8c2ddde, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.068245] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.068724] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.068724] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.069071] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.069300] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.073880] env[66641]: INFO nova.compute.manager [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Terminating instance [ 656.080081] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.103617] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145631, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.264075] env[66641]: DEBUG nova.scheduler.client.report [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 656.275622] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 656.522464] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145632, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.542221] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217d37c-8ec7-a7ad-a445-c119c8c2ddde, 'name': SearchDatastore_Task, 'duration_secs': 0.017241} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.543125] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b70224d-4a1f-4b3e-9078-67ae752b307f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.550604] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 656.550604] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52725df2-816f-6682-f192-45a4d1c8d414" [ 656.550604] env[66641]: _type = "Task" [ 656.550604] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.562208] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52725df2-816f-6682-f192-45a4d1c8d414, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.587733] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145631, 'name': ReconfigVM_Task, 'duration_secs': 0.860594} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 656.588065] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Reconfigured VM instance instance-00000009 to attach disk [datastore2] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 656.588695] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4f61cb49-f41e-4ee8-aac4-906c81052967 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.593843] env[66641]: DEBUG nova.compute.manager [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 656.593972] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 656.594832] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cae0f47-d4ff-4106-9259-7a2851fa67c9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.599403] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 656.599403] env[66641]: value = "task-5145634" [ 656.599403] env[66641]: _type = "Task" [ 656.599403] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.605782] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 656.606591] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-eaf3c784-4ab4-4ba7-b226-b3188478e096 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.612653] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145634, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.614334] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 656.614334] env[66641]: value = "task-5145635" [ 656.614334] env[66641]: _type = "Task" [ 656.614334] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 656.626902] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 656.768541] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.941s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 656.772143] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.250s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 656.774789] env[66641]: INFO nova.compute.claims [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.810024] env[66641]: INFO nova.scheduler.client.report [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Deleted allocations for instance ccf46c5c-1c79-4672-ad9f-ea61042097d5 [ 656.846792] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 656.881650] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 656.882076] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 656.882233] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 656.882480] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 656.882679] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 656.882966] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 656.883166] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.883356] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 656.883550] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 656.883795] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 656.884039] env[66641]: DEBUG nova.virt.hardware [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 656.885046] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43794d5d-e936-476d-87e2-e665cad97e2b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.894716] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e4513f0-51c6-4326-93e7-f457f38b87e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.023116] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145632, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.788405} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.023739] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 657.023955] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 657.024236] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c8069f49-30cf-4cec-bd18-fc2fb209ad86 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.031806] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 657.031806] env[66641]: value = "task-5145636" [ 657.031806] env[66641]: _type = "Task" [ 657.031806] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.042874] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145636, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.062906] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52725df2-816f-6682-f192-45a4d1c8d414, 'name': SearchDatastore_Task, 'duration_secs': 0.068337} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.063114] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 657.063422] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 6b19204d-032c-410b-bb84-1af23d142edc/6b19204d-032c-410b-bb84-1af23d142edc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 657.063697] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b9d977c-6cf2-48ba-80f1-46b7477bb29f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.072352] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 657.072352] env[66641]: value = "task-5145637" [ 657.072352] env[66641]: _type = "Task" [ 657.072352] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.082521] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145637, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.114023] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145634, 'name': Rename_Task, 'duration_secs': 0.257474} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.114564] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 657.115607] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b7fbb22d-d4a5-4699-ab4e-4c96e93188b4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.128133] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145635, 'name': PowerOffVM_Task, 'duration_secs': 0.253138} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.129955] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 657.130462] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 657.131244] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 657.131244] env[66641]: value = "task-5145638" [ 657.131244] env[66641]: _type = "Task" [ 657.131244] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.132183] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50de6e8a-6d01-421c-a49e-597d82d18d76 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.144787] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145638, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.231774] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 657.232127] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 657.232351] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Deleting the datastore file [datastore1] 47aaed25-542b-4ceb-9adf-6a4953c8c95d {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 657.232750] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5cc4ec3-decc-4bbf-bef2-4e3b38029cda {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.244028] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for the task: (returnval){ [ 657.244028] env[66641]: value = "task-5145640" [ 657.244028] env[66641]: _type = "Task" [ 657.244028] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.248337] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.248701] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.259036] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.279929] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.508s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.280938] env[66641]: DEBUG nova.compute.utils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Instance 7fa45876-59c2-4fa2-978b-46067931e091 could not be found. {{(pid=66641) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 657.282508] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.948s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.284590] env[66641]: INFO nova.compute.claims [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.288148] env[66641]: DEBUG nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Instance disappeared during build. {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2556}} [ 657.288796] env[66641]: DEBUG nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Unplugging VIFs for instance {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3050}} [ 657.288796] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "refresh_cache-7fa45876-59c2-4fa2-978b-46067931e091" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.288796] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquired lock "refresh_cache-7fa45876-59c2-4fa2-978b-46067931e091" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.289077] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 657.319628] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.320215] env[66641]: WARNING openstack [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.330652] env[66641]: DEBUG oslo_concurrency.lockutils [None req-807ac2e7-6145-433f-9d42-19b1dff14c61 tempest-ServerDiagnosticsTest-2114745596 tempest-ServerDiagnosticsTest-2114745596-project-member] Lock "ccf46c5c-1c79-4672-ad9f-ea61042097d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.092s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.371494] env[66641]: DEBUG nova.compute.manager [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Received event network-vif-plugged-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 657.371753] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Acquiring lock "6b19204d-032c-410b-bb84-1af23d142edc-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 657.371988] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Lock "6b19204d-032c-410b-bb84-1af23d142edc-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 657.372101] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Lock "6b19204d-032c-410b-bb84-1af23d142edc-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 657.372325] env[66641]: DEBUG nova.compute.manager [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] No waiting events found dispatching network-vif-plugged-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 657.372460] env[66641]: WARNING nova.compute.manager [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Received unexpected event network-vif-plugged-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f for instance with vm_state building and task_state spawning. [ 657.372682] env[66641]: DEBUG nova.compute.manager [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Received event network-changed-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 657.372847] env[66641]: DEBUG nova.compute.manager [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Refreshing instance network info cache due to event network-changed-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 657.373063] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Acquiring lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.374756] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Acquired lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 657.374756] env[66641]: DEBUG nova.network.neutron [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Refreshing network info cache for port 9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 657.523721] env[66641]: DEBUG nova.network.neutron [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Updated VIF entry in instance network info cache for port aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 657.524159] env[66641]: DEBUG nova.network.neutron [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Updating instance_info_cache with network_info: [{"id": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "address": "fa:16:3e:d8:c3:ea", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaa8a5251-c9", "ovs_interfaceid": "aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 657.555708] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145636, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082798} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.556029] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 657.558051] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0558cffe-37a9-430b-8a8e-b0f38e7d1311 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.584680] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 657.589158] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9731b9b6-2a3f-4bef-800c-289099704ff3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.616015] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145637, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.620030] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 657.620030] env[66641]: value = "task-5145641" [ 657.620030] env[66641]: _type = "Task" [ 657.620030] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 657.630246] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145641, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.646171] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145638, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 657.759159] env[66641]: DEBUG oslo_vmware.api [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Task: {'id': task-5145640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.459133} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 657.759159] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 657.759159] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 657.759159] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 657.759159] env[66641]: INFO nova.compute.manager [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Took 1.16 seconds to destroy the instance on the hypervisor. [ 657.759539] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 657.759539] env[66641]: DEBUG nova.compute.manager [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 657.759539] env[66641]: DEBUG nova.network.neutron [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 657.760127] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.761629] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.792304] env[66641]: DEBUG nova.compute.utils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Can not refresh info_cache because instance was not found {{(pid=66641) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 657.793670] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.794775] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 657.801128] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 657.834232] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Successfully updated port: d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 657.878956] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 657.879399] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.010115] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.010405] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.029921] env[66641]: DEBUG oslo_concurrency.lockutils [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] Releasing lock "refresh_cache-fe143d94-57ce-46d4-85e3-80d84af66dfc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.030343] env[66641]: DEBUG nova.compute.manager [req-cc783220-6bce-4cc7-b709-9a729ecd722a req-ae17e74f-41df-400a-b5ef-6ec6bcdb7228 service nova] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Received event network-vif-deleted-3e466382-fc5f-4ac9-90dc-05e1d9f93c48 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 658.087015] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145637, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.595949} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.087509] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 6b19204d-032c-410b-bb84-1af23d142edc/6b19204d-032c-410b-bb84-1af23d142edc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 658.087509] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 658.088133] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-44ebb846-75c1-4e50-9dc3-fc7f4c95fb1f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.096487] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 658.096487] env[66641]: value = "task-5145642" [ 658.096487] env[66641]: _type = "Task" [ 658.096487] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.107706] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.131293] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145641, 'name': ReconfigVM_Task, 'duration_secs': 0.408451} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.131618] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 658.132289] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bbfe3377-1e45-4d38-940b-20e31bf77a88 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.140860] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 658.140860] env[66641]: value = "task-5145643" [ 658.140860] env[66641]: _type = "Task" [ 658.140860] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.152345] env[66641]: DEBUG oslo_vmware.api [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145638, 'name': PowerOnVM_Task, 'duration_secs': 0.601104} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.152941] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 658.153202] env[66641]: INFO nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Took 12.73 seconds to spawn the instance on the hypervisor. [ 658.153463] env[66641]: DEBUG nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 658.154525] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a73aa9-5e5a-4048-b63d-16eabecba2d8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.160751] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145643, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.323783] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.342088] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.342088] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquired lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 658.342088] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 658.352346] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.352346] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.592629] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b764364-e5bb-469e-9a89-24ca87bfeb24 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.606521] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4095f5d9-94de-4ffa-8b06-d4d21ba9d9b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.613260] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.099578} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.613983] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 658.614838] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f67401c-c83f-499c-a65e-287fb67269ff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.658452] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28218e29-2ef4-457f-b7e7-01afdbd7c3a6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.681553] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] 6b19204d-032c-410b-bb84-1af23d142edc/6b19204d-032c-410b-bb84-1af23d142edc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 658.685683] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5eb443ae-9ac7-4696-8686-324e48ef9ca3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.711844] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145643, 'name': Rename_Task, 'duration_secs': 0.173649} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 658.711844] env[66641]: INFO nova.compute.manager [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Took 17.64 seconds to build instance. [ 658.713108] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 658.714687] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6a1835-ecd8-42d4-9575-c95b94a7cbe3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.721851] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e09adf3a-5c10-45d5-8f8d-02caa36b7b47 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.724260] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 658.724260] env[66641]: value = "task-5145645" [ 658.724260] env[66641]: _type = "Task" [ 658.724260] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.738590] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 658.738590] env[66641]: value = "task-5145646" [ 658.738590] env[66641]: _type = "Task" [ 658.738590] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 658.739579] env[66641]: DEBUG nova.compute.provider_tree [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.748725] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145645, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.754538] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145646, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 658.787528] env[66641]: DEBUG nova.network.neutron [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 658.828886] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Releasing lock "refresh_cache-7fa45876-59c2-4fa2-978b-46067931e091" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 658.830588] env[66641]: DEBUG nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3073}} [ 658.830854] env[66641]: DEBUG nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 658.831109] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 658.832244] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.832841] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.851450] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.851970] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 658.858360] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 658.893887] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 658.896139] env[66641]: WARNING openstack [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.027701] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 659.027701] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.027701] env[66641]: WARNING openstack [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.135958] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.136401] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.176169] env[66641]: DEBUG nova.network.neutron [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Updated VIF entry in instance network info cache for port 9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 659.176596] env[66641]: DEBUG nova.network.neutron [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Updating instance_info_cache with network_info: [{"id": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "address": "fa:16:3e:bf:1b:08", "network": {"id": "0a7d6c39-f10c-4cd3-9740-fc9c04d9962f", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-623599665-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "123276c617f54856923563091fd1842c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "29945b49-d2b8-449d-9531-437917f49839", "external-id": "nsx-vlan-transportzone-787", "segmentation_id": 787, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c5aba0e-5f", "ovs_interfaceid": "9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.221473] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8803a4af-3064-47c2-8ced-f6f03330f3ae tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.159s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.234536] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145645, 'name': ReconfigVM_Task, 'duration_secs': 0.311281} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.235329] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Reconfigured VM instance instance-0000000a to attach disk [datastore2] 6b19204d-032c-410b-bb84-1af23d142edc/6b19204d-032c-410b-bb84-1af23d142edc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 659.235877] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0b6b7afe-e9c6-4c44-aa86-78de8d8c7d93 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.241047] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 659.241481] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 659.250162] env[66641]: DEBUG nova.scheduler.client.report [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 659.263975] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 659.263975] env[66641]: value = "task-5145647" [ 659.263975] env[66641]: _type = "Task" [ 659.263975] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.271441] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145646, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.278264] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145647, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.290517] env[66641]: INFO nova.compute.manager [-] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Took 1.53 seconds to deallocate network for instance. [ 659.368599] env[66641]: DEBUG nova.network.neutron [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Updating instance_info_cache with network_info: [{"id": "d3f790d9-7633-46ba-842e-725e9680d0ea", "address": "fa:16:3e:a7:27:b6", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f790d9-76", "ovs_interfaceid": "d3f790d9-7633-46ba-842e-725e9680d0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.547522] env[66641]: DEBUG nova.network.neutron [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 659.679366] env[66641]: DEBUG oslo_concurrency.lockutils [req-570703e5-000f-4ee4-a2d2-3e92c71451d1 req-3dc77c07-b506-4a00-8af4-517eb4565176 service nova] Releasing lock "refresh_cache-6b19204d-032c-410b-bb84-1af23d142edc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.759805] env[66641]: DEBUG oslo_vmware.api [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145646, 'name': PowerOnVM_Task, 'duration_secs': 0.761774} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.760406] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 659.760599] env[66641]: INFO nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Took 7.18 seconds to spawn the instance on the hypervisor. [ 659.760815] env[66641]: DEBUG nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 659.761989] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88774c38-f20e-4550-b44d-0d255e2ed723 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.768596] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.768596] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 659.772062] env[66641]: DEBUG nova.compute.manager [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Received event network-vif-plugged-d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 659.772273] env[66641]: DEBUG oslo_concurrency.lockutils [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] Acquiring lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.772469] env[66641]: DEBUG oslo_concurrency.lockutils [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.772623] env[66641]: DEBUG oslo_concurrency.lockutils [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 659.772931] env[66641]: DEBUG nova.compute.manager [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] No waiting events found dispatching network-vif-plugged-d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 659.772931] env[66641]: WARNING nova.compute.manager [req-ffcca38d-e455-477c-8753-df1352ceeef0 req-17f79b44-bc4e-4776-9232-d6039e64a718 service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Received unexpected event network-vif-plugged-d3f790d9-7633-46ba-842e-725e9680d0ea for instance with vm_state building and task_state spawning. [ 659.773468] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.164s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 659.773678] env[66641]: DEBUG nova.objects.instance [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lazy-loading 'resources' on Instance uuid 4ebc0f2b-bb79-4941-b8f2-082560c80f8b {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 659.792009] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145647, 'name': Rename_Task, 'duration_secs': 0.156676} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 659.792468] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 659.792930] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afffdbb2-f6f0-4ba9-bd62-145211b7721a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.799120] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 659.806050] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 659.806050] env[66641]: value = "task-5145648" [ 659.806050] env[66641]: _type = "Task" [ 659.806050] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.822677] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145648, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 659.875816] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Releasing lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 659.875995] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Instance network_info: |[{"id": "d3f790d9-7633-46ba-842e-725e9680d0ea", "address": "fa:16:3e:a7:27:b6", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f790d9-76", "ovs_interfaceid": "d3f790d9-7633-46ba-842e-725e9680d0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 659.876383] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:27:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3f790d9-7633-46ba-842e-725e9680d0ea', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 659.886409] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Creating folder: Project (91b733f620ed4947975d30e02ea73e24). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.887691] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9ae68409-15cf-4a56-b48c-e86bb9b2fc4f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.899114] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Created folder: Project (91b733f620ed4947975d30e02ea73e24) in parent group-v1000566. [ 659.899347] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Creating folder: Instances. Parent ref: group-v1000603. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 659.899610] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f79614df-f1b8-4a52-987e-d339c089a362 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.909639] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Created folder: Instances in parent group-v1000603. [ 659.909940] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 659.910192] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 659.910748] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a0ef345-f8bf-4537-8810-57baea2246c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.933765] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 659.933765] env[66641]: value = "task-5145651" [ 659.933765] env[66641]: _type = "Task" [ 659.933765] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 659.944997] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145651, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.051904] env[66641]: INFO nova.compute.manager [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 7fa45876-59c2-4fa2-978b-46067931e091] Took 1.22 seconds to deallocate network for instance. [ 660.276034] env[66641]: DEBUG nova.compute.utils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 660.277978] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 660.277978] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 660.278330] env[66641]: WARNING neutronclient.v2_0.client [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.278671] env[66641]: WARNING neutronclient.v2_0.client [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 660.279357] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.279774] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.303364] env[66641]: INFO nova.compute.manager [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Took 15.46 seconds to build instance. [ 660.318920] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145648, 'name': PowerOnVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.444135] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145651, 'name': CreateVM_Task, 'duration_secs': 0.475531} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.444328] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 660.445095] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 660.445462] env[66641]: WARNING openstack [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 660.451800] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.452109] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.452543] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 660.457801] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ee8832d-0097-403a-a1ed-eaa84bae4852 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.464183] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 660.464183] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa348a-b3eb-7264-4f26-d6ddb838f424" [ 660.464183] env[66641]: _type = "Task" [ 660.464183] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 660.480452] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa348a-b3eb-7264-4f26-d6ddb838f424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 660.584178] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-606fd5a1-8044-40c1-9f36-6c7fe6a35ea9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.594275] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec141abf-63af-4636-9a3e-3018bdd0359b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.602781] env[66641]: DEBUG nova.policy [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf590217db704af2acb3d078ce6d789f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17792883c6504d0f9f91fb6beb25b087', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 660.634850] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ed4fb2-2fa2-45e5-9ef6-8ca04f9c29fa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.642974] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1b069a-f407-4951-9f86-e1dc65831dde {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.661065] env[66641]: DEBUG nova.compute.provider_tree [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.788739] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 660.806643] env[66641]: DEBUG oslo_concurrency.lockutils [None req-85793071-8d58-45b1-a1ac-ecec33fa64e8 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.977s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 660.816488] env[66641]: DEBUG oslo_vmware.api [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145648, 'name': PowerOnVM_Task, 'duration_secs': 0.740429} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.816857] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 660.816999] env[66641]: INFO nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Took 10.61 seconds to spawn the instance on the hypervisor. [ 660.817513] env[66641]: DEBUG nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 660.818170] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241ad542-f757-49f2-8d3a-a484c1e7cdb5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.975930] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa348a-b3eb-7264-4f26-d6ddb838f424, 'name': SearchDatastore_Task, 'duration_secs': 0.018} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 660.976327] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 660.976416] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 660.976627] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.976771] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 660.976946] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 660.977236] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6a55aef0-4979-4d5a-ba05-aab5fc9cb207 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.988401] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 660.988648] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 660.989372] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-864cf74d-8b78-4703-a711-e9eb52eeb61a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.997605] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 660.997605] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52467be6-4a1a-1da8-0461-f110720720fe" [ 660.997605] env[66641]: _type = "Task" [ 660.997605] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.006796] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52467be6-4a1a-1da8-0461-f110720720fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.079773] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8f77719f-14c8-496a-954f-e0ffb13235b1 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "7fa45876-59c2-4fa2-978b-46067931e091" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 9.605s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.163779] env[66641]: DEBUG nova.scheduler.client.report [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 661.320310] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Successfully created port: 4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 661.336801] env[66641]: INFO nova.compute.manager [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Took 17.11 seconds to build instance. [ 661.519774] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52467be6-4a1a-1da8-0461-f110720720fe, 'name': SearchDatastore_Task, 'duration_secs': 0.017604} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 661.521229] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-104c8ab6-e68b-4847-8eae-4a521a80cda5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.529042] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 661.529042] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52754d26-03e0-ae5b-9d88-6e1d32efcf8a" [ 661.529042] env[66641]: _type = "Task" [ 661.529042] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 661.541101] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52754d26-03e0-ae5b-9d88-6e1d32efcf8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 661.670123] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.896s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.674313] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.359s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 661.710356] env[66641]: INFO nova.scheduler.client.report [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Deleted allocations for instance 4ebc0f2b-bb79-4941-b8f2-082560c80f8b [ 661.801539] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 661.826467] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 661.826467] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 661.827238] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 661.827943] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 661.828424] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 661.828767] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 661.829169] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.830993] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 661.830993] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 661.830993] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 661.830993] env[66641]: DEBUG nova.virt.hardware [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 661.831389] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43830bb8-eba7-4290-912c-bb201f67a9e8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.842024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ab4c9f92-2296-4f79-bef4-c6e766801b80 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.643s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 661.846039] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d15e6e98-b14b-4c29-b8e7-bd900baca858 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.040568] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52754d26-03e0-ae5b-9d88-6e1d32efcf8a, 'name': SearchDatastore_Task, 'duration_secs': 0.015211} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 662.040827] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 662.041136] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] e7bfb26e-b326-4a39-8f20-79dbbfb74eb8/e7bfb26e-b326-4a39-8f20-79dbbfb74eb8.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 662.041404] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5caf0002-5bb0-4fe6-b18e-6059de9df0b4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.048894] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 662.048894] env[66641]: value = "task-5145653" [ 662.048894] env[66641]: _type = "Task" [ 662.048894] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.057834] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145653, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.183429] env[66641]: INFO nova.compute.claims [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.219720] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0ccf92-12dc-49c4-9809-4139da01c83c tempest-ServerDiagnosticsV248Test-734459867 tempest-ServerDiagnosticsV248Test-734459867-project-member] Lock "4ebc0f2b-bb79-4941-b8f2-082560c80f8b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.377s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.277230] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "6b19204d-032c-410b-bb84-1af23d142edc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.277610] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.277912] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "6b19204d-032c-410b-bb84-1af23d142edc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 662.278205] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 662.278466] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 662.281423] env[66641]: INFO nova.compute.manager [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Terminating instance [ 662.561571] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145653, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.690024] env[66641]: INFO nova.compute.resource_tracker [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating resource usage from migration 1ed74871-41c7-4c89-bbf4-b55c237571e3 [ 662.786945] env[66641]: DEBUG nova.compute.manager [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 662.787198] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 662.788100] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c982dd-4933-4057-a115-1c73690ac30d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.799968] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 662.800284] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-505b9f05-6670-48df-9730-435f3a503fb1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.808131] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 662.808131] env[66641]: value = "task-5145655" [ 662.808131] env[66641]: _type = "Task" [ 662.808131] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 662.823619] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145655, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 662.891710] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9910e11-cf8b-4650-88de-9aefde0bdeb1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.900292] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12833a0d-c3f9-42a9-9b68-821de0e38939 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.932504] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e7e5bb-bd33-4a63-8e07-5fd947ac2c2f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.940632] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-690d001c-5851-48de-96f1-cf37a654a567 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.954642] env[66641]: DEBUG nova.compute.provider_tree [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.062799] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145653, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567263} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.063130] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] e7bfb26e-b326-4a39-8f20-79dbbfb74eb8/e7bfb26e-b326-4a39-8f20-79dbbfb74eb8.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 663.063499] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 663.063835] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-95f8afe7-469a-4eea-9524-304bc80234cb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.072150] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 663.072150] env[66641]: value = "task-5145656" [ 663.072150] env[66641]: _type = "Task" [ 663.072150] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.084333] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.203832] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Successfully updated port: 4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 663.328230] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145655, 'name': PowerOffVM_Task, 'duration_secs': 0.219434} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.328503] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 663.328664] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 663.328951] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de46386a-43a5-480a-94a9-12cdaf8ba4e3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.401496] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 663.401722] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 663.402673] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Deleting the datastore file [datastore2] 6b19204d-032c-410b-bb84-1af23d142edc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 663.402673] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2511370e-9f35-4aa4-bd1c-41776c78976d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.410900] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for the task: (returnval){ [ 663.410900] env[66641]: value = "task-5145658" [ 663.410900] env[66641]: _type = "Task" [ 663.410900] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.423541] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.461424] env[66641]: DEBUG nova.scheduler.client.report [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 663.582718] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074656} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 663.582984] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 663.584284] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f800eef3-3417-474a-94cd-e0b45d1f7bec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.614496] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Reconfiguring VM instance instance-0000000c to attach disk [datastore1] e7bfb26e-b326-4a39-8f20-79dbbfb74eb8/e7bfb26e-b326-4a39-8f20-79dbbfb74eb8.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 663.614496] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ab023838-c259-4270-9764-0d0ecb261a8b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.637475] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 663.637475] env[66641]: value = "task-5145659" [ 663.637475] env[66641]: _type = "Task" [ 663.637475] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 663.646527] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145659, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.707458] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.707685] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.707860] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 663.929105] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145658, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 663.972624] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.297s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 663.972624] env[66641]: INFO nova.compute.manager [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Migrating [ 663.972624] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.972624] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "compute-rpcapi-router" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 663.974659] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.624s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 663.974954] env[66641]: DEBUG nova.objects.instance [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lazy-loading 'resources' on Instance uuid d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 664.151745] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145659, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.213454] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.213454] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.234662] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 664.442472] env[66641]: DEBUG oslo_vmware.api [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Task: {'id': task-5145658, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.56579} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.442472] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 664.442472] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 664.442472] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.442749] env[66641]: INFO nova.compute.manager [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Took 1.66 seconds to destroy the instance on the hypervisor. [ 664.443122] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 664.443122] env[66641]: DEBUG nova.compute.manager [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 664.443245] env[66641]: DEBUG nova.network.neutron [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 664.446222] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.446222] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.480024] env[66641]: INFO nova.compute.rpcapi [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Automatically selected compute RPC version 6.4 from minimum service version 70 [ 664.480024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "compute-rpcapi-router" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 664.609260] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.609905] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.653679] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 664.654012] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 664.671620] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145659, 'name': ReconfigVM_Task, 'duration_secs': 0.973702} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 664.672360] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Reconfigured VM instance instance-0000000c to attach disk [datastore1] e7bfb26e-b326-4a39-8f20-79dbbfb74eb8/e7bfb26e-b326-4a39-8f20-79dbbfb74eb8.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 664.672964] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-077cac15-c7d7-40cd-a02b-c0e68a9a5dd8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.681879] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 664.681879] env[66641]: value = "task-5145661" [ 664.681879] env[66641]: _type = "Task" [ 664.681879] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 664.693169] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145661, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 664.805479] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f659e7-60c0-46f8-a68a-d1f18b9676e7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.818184] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b43b4a-bbe0-4893-bbc2-b9f0a0de3c68 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.869425] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3174e238-c3dd-4c3b-b045-568c6bca1ed6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.879443] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a69e589-ea69-4ab1-b081-14ae0293385d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.896079] env[66641]: DEBUG nova.compute.provider_tree [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.021104] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.021104] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 665.021104] env[66641]: DEBUG nova.network.neutron [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 665.062224] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.062224] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.194924] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145661, 'name': Rename_Task, 'duration_secs': 0.205362} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 665.195635] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 665.196356] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b28acc5f-b4aa-4713-912f-59151d6a53b5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.203808] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 665.203808] env[66641]: value = "task-5145662" [ 665.203808] env[66641]: _type = "Task" [ 665.203808] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 665.213947] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145662, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.401753] env[66641]: DEBUG nova.scheduler.client.report [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 665.496123] env[66641]: DEBUG nova.network.neutron [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.525960] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 665.526426] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 665.717467] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145662, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 665.720928] env[66641]: INFO nova.compute.manager [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Rebuilding instance [ 665.731658] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 665.732323] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.766193] env[66641]: DEBUG nova.network.neutron [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Updating instance_info_cache with network_info: [{"id": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "address": "fa:16:3e:f8:4e:ac", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c64a4dc-a0", "ovs_interfaceid": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 665.817489] env[66641]: DEBUG nova.compute.manager [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 665.818605] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d941b669-90ce-420e-8fee-7949c4bf7674 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.910685] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 665.914733] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.976s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 665.916423] env[66641]: INFO nova.compute.claims [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 666.002159] env[66641]: INFO nova.compute.manager [-] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Took 1.56 seconds to deallocate network for instance. [ 666.217753] env[66641]: DEBUG oslo_vmware.api [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145662, 'name': PowerOnVM_Task, 'duration_secs': 0.703564} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 666.222477] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 666.222725] env[66641]: INFO nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Took 9.38 seconds to spawn the instance on the hypervisor. [ 666.223190] env[66641]: DEBUG nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 666.224011] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b1a578-26d6-4de0-ad0b-4d9486c2d2c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.236263] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 666.276423] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 666.278039] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Instance network_info: |[{"id": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "address": "fa:16:3e:f8:4e:ac", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c64a4dc-a0", "ovs_interfaceid": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 666.278235] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f8:4e:ac', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4c64a4dc-a032-4361-b383-ea4b1d7c3728', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 666.285788] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating folder: Project (17792883c6504d0f9f91fb6beb25b087). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.286154] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-75484763-cb00-4a75-bb48-4725437a0295 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.299609] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Created folder: Project (17792883c6504d0f9f91fb6beb25b087) in parent group-v1000566. [ 666.299844] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating folder: Instances. Parent ref: group-v1000607. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 666.300112] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-119e6901-27e1-482b-8a6a-84ad1503864a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.316208] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Created folder: Instances in parent group-v1000607. [ 666.316208] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 666.316208] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 666.316208] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-58722c0f-e81e-474e-b7ee-25ea39b77ff9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.344641] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 666.344641] env[66641]: value = "task-5145665" [ 666.344641] env[66641]: _type = "Task" [ 666.344641] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.356117] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145665, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.375264] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 666.375646] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 666.443426] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5cb0f66e-67fc-483e-8270-8b6c54b591a9 tempest-ServersTestJSON-1435660074 tempest-ServersTestJSON-1435660074-project-member] Lock "d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.633s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 666.512056] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.771356] env[66641]: INFO nova.compute.manager [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Took 19.03 seconds to build instance. [ 666.775399] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 666.838198] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 666.838402] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5e45b6f-4548-4539-a5a6-00143f821d96 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.849420] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 666.849420] env[66641]: value = "task-5145667" [ 666.849420] env[66641]: _type = "Task" [ 666.849420] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 666.858194] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145665, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 666.864319] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145667, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.185745] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64bb01dc-619f-4035-baf6-cb29eb3e8380 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.196870] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eadcf2-1877-4830-a31c-07dcf78a3670 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.238092] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95f0e581-97ad-47ae-a923-a8f9588b1210 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.249208] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69259a2e-1d75-4149-9f85-4cbf57ca5506 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.273385] env[66641]: DEBUG nova.compute.provider_tree [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.276623] env[66641]: DEBUG oslo_concurrency.lockutils [None req-164c33bf-1a74-4372-b2e5-ccac53fc38d1 tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.540s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 667.281577] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.281577] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.357373] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145665, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.363606] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145667, 'name': PowerOffVM_Task, 'duration_secs': 0.302032} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.363916] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 667.364175] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.364983] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4ecbda-1ab1-4114-842b-6442d5e22a34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.373762] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 667.374091] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6446006b-4788-49bd-ae11-e1eac64b00ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.401769] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 667.401967] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 667.402199] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Deleting the datastore file [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 667.402534] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c5655dea-8069-4cca-94c9-dee9356c0ec1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.414586] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 667.414586] env[66641]: value = "task-5145669" [ 667.414586] env[66641]: _type = "Task" [ 667.414586] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.423981] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145669, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.779890] env[66641]: DEBUG nova.scheduler.client.report [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 667.856846] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145665, 'name': CreateVM_Task, 'duration_secs': 1.39329} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.857811] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 667.857811] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 667.858092] env[66641]: WARNING openstack [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 667.863489] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.863651] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 667.864272] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 667.864555] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b191afe3-37f2-46b0-9573-4cbe71413ae3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.871080] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 667.871080] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbbedd-6338-0177-8167-9186cd95f720" [ 667.871080] env[66641]: _type = "Task" [ 667.871080] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 667.880625] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbbedd-6338-0177-8167-9186cd95f720, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 667.926560] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145669, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159532} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 667.926796] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 667.927430] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 667.927642] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 668.290019] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.372s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 668.290019] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 668.291134] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.302s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 668.296199] env[66641]: INFO nova.compute.claims [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 668.381930] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbbedd-6338-0177-8167-9186cd95f720, 'name': SearchDatastore_Task, 'duration_secs': 0.033847} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.382156] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 668.382448] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 668.382692] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.382833] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 668.383013] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 668.383293] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d8a928e-8c13-4ac5-93ec-89f5a447cf3a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.393360] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 668.393525] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 668.394365] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5df45086-11ce-45f9-8cff-938922e3876e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.400197] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 668.400197] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5241f264-ce0a-8f2e-5c0d-65121735ee1c" [ 668.400197] env[66641]: _type = "Task" [ 668.400197] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.409647] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5241f264-ce0a-8f2e-5c0d-65121735ee1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.803211] env[66641]: DEBUG nova.compute.utils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 668.810221] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 668.811250] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 668.811250] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.812525] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 668.813433] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 668.813528] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 668.913316] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5241f264-ce0a-8f2e-5c0d-65121735ee1c, 'name': SearchDatastore_Task, 'duration_secs': 0.012271} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 668.914206] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8b2572a-0d52-45e4-ac8c-072731856de2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.921789] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 668.921789] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e98fb3-7228-d8bd-0c76-cca9cb7b13ea" [ 668.921789] env[66641]: _type = "Task" [ 668.921789] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 668.933108] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e98fb3-7228-d8bd-0c76-cca9cb7b13ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 668.950851] env[66641]: DEBUG nova.network.neutron [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 668.965165] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 668.965890] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 668.965890] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 668.965890] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 668.966079] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 668.966255] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 668.967612] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.967612] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 668.967612] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 668.967612] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 668.967612] env[66641]: DEBUG nova.virt.hardware [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 668.968195] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68092aa7-ba64-433f-a168-da72f0cabf36 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.980211] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b542a8d6-4f6b-44c0-ab08-81a9f37b7ffd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.995798] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 669.001801] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 669.002213] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 669.002504] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-df6c6487-3c88-4373-9fbd-501b5d08509e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.021942] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 669.021942] env[66641]: value = "task-5145670" [ 669.021942] env[66641]: _type = "Task" [ 669.021942] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.032889] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145670, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.069739] env[66641]: DEBUG nova.compute.manager [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Received event network-changed-d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 669.069739] env[66641]: DEBUG nova.compute.manager [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Refreshing instance network info cache due to event network-changed-d3f790d9-7633-46ba-842e-725e9680d0ea. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 669.069739] env[66641]: DEBUG oslo_concurrency.lockutils [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Acquiring lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.069739] env[66641]: DEBUG oslo_concurrency.lockutils [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Acquired lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 669.069739] env[66641]: DEBUG nova.network.neutron [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Refreshing network info cache for port d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 669.169923] env[66641]: DEBUG nova.policy [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7279b628bf1646d8bd27416d1f52cfdb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196b2ed18f7142a18f9093869dadb46e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 669.308712] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 669.434090] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e98fb3-7228-d8bd-0c76-cca9cb7b13ea, 'name': SearchDatastore_Task, 'duration_secs': 0.017307} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 669.434459] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.434739] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 695d1824-cf34-46c1-b623-b474d2f1f8ef/695d1824-cf34-46c1-b623-b474d2f1f8ef.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 669.435051] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-787ff317-fbdc-4e94-840d-175c5399277c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.450923] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 669.450923] env[66641]: value = "task-5145671" [ 669.450923] env[66641]: _type = "Task" [ 669.450923] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.456024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 669.463395] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145671, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.540031] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145670, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.575718] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 669.576327] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 669.670996] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c027d1b4-9627-460a-9d0e-035da4853197 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.679669] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c14175-3032-4d75-a9d2-76d394a99e16 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.715780] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5c9f7a-e70e-4278-b108-2b7b1f09e934 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.724991] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42c2c355-d9ae-4140-bf42-743a09c084aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.742656] env[66641]: DEBUG nova.compute.provider_tree [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 669.955280] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Successfully created port: 2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 669.970185] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145671, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.036979] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145670, 'name': CreateVM_Task, 'duration_secs': 0.554327} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.037210] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 670.037662] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.037839] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.038269] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 670.038569] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9914103c-b8ba-45aa-bcba-d6f810a3b666 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.045243] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 670.045243] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a385e0-eae0-4555-bcba-429492244072" [ 670.045243] env[66641]: _type = "Task" [ 670.045243] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.055218] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a385e0-eae0-4555-bcba-429492244072, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.247205] env[66641]: DEBUG nova.scheduler.client.report [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 670.322567] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 670.354304] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 670.354615] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 670.354779] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 670.355026] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 670.355173] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 670.355327] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 670.355529] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 670.355711] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 670.355882] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 670.356063] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 670.356212] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 670.357062] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05909501-89a3-41d9-8a01-9692af3a38d8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.367425] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd2f8b65-57bb-4d9d-8307-0105a1f365f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.469407] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145671, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689493} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.469507] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 695d1824-cf34-46c1-b623-b474d2f1f8ef/695d1824-cf34-46c1-b623-b474d2f1f8ef.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 670.469675] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 670.470041] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cd23f680-1b85-4c35-8009-76449614e739 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.498786] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 670.498786] env[66641]: value = "task-5145672" [ 670.498786] env[66641]: _type = "Task" [ 670.498786] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.498786] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145672, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.561414] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a385e0-eae0-4555-bcba-429492244072, 'name': SearchDatastore_Task, 'duration_secs': 0.053501} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.561414] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 670.561656] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 670.561887] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.562182] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 670.562514] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 670.562882] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0bf5e2d6-2f50-448b-a7a9-f0b3d6d035e7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.575114] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 670.575389] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 670.576225] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19ac2a1e-a899-4f64-b926-d2aeb6092c89 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.586912] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 670.586912] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b64bfd-5069-3165-ecba-451ec287a9c2" [ 670.586912] env[66641]: _type = "Task" [ 670.586912] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.600034] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b64bfd-5069-3165-ecba-451ec287a9c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.754273] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 670.754837] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 670.757698] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.678s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 670.759113] env[66641]: INFO nova.compute.claims [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 670.982249] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 670.983705] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 670.996901] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42bfb87f-d166-49ce-925f-82cde7bba599 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.005306] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145672, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110851} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.023385] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 671.024903] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 0 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 671.028221] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af717b27-5b43-4da9-835a-ba38d32ed34a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.055026] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Reconfiguring VM instance instance-0000000e to attach disk [datastore1] 695d1824-cf34-46c1-b623-b474d2f1f8ef/695d1824-cf34-46c1-b623-b474d2f1f8ef.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 671.055336] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-55dd31ae-b3c7-4dcd-8670-abd571dc8506 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.076349] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 671.076349] env[66641]: value = "task-5145673" [ 671.076349] env[66641]: _type = "Task" [ 671.076349] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.085735] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145673, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.097922] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b64bfd-5069-3165-ecba-451ec287a9c2, 'name': SearchDatastore_Task, 'duration_secs': 0.012557} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.098758] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d87f0c6-e50d-48af-8e11-d33ed652f079 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.105485] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 671.105485] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c38a38-bfe7-d6d9-b741-26056616653d" [ 671.105485] env[66641]: _type = "Task" [ 671.105485] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.114448] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c38a38-bfe7-d6d9-b741-26056616653d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.267753] env[66641]: DEBUG nova.compute.utils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 671.269958] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 671.270096] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 671.270408] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.270756] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 671.271563] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 671.272037] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 671.535428] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 671.535428] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60c5289a-33ff-4639-b051-4bf181875cf5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.544575] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 671.544575] env[66641]: value = "task-5145674" [ 671.544575] env[66641]: _type = "Task" [ 671.544575] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.557539] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145674, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.590824] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145673, 'name': ReconfigVM_Task, 'duration_secs': 0.280302} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.590824] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Reconfigured VM instance instance-0000000e to attach disk [datastore1] 695d1824-cf34-46c1-b623-b474d2f1f8ef/695d1824-cf34-46c1-b623-b474d2f1f8ef.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 671.590976] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a8e0e8b3-4910-42d1-b9f4-21814eaf42a0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.599488] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 671.599488] env[66641]: value = "task-5145675" [ 671.599488] env[66641]: _type = "Task" [ 671.599488] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.622556] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145675, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.622556] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c38a38-bfe7-d6d9-b741-26056616653d, 'name': SearchDatastore_Task, 'duration_secs': 0.018478} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.622871] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 671.623212] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 671.623898] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4be51a19-cb9d-47f4-9360-acded8b3443e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.631220] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 671.631220] env[66641]: value = "task-5145676" [ 671.631220] env[66641]: _type = "Task" [ 671.631220] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.640594] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145676, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.781725] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 671.824172] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Successfully updated port: 2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 671.870459] env[66641]: DEBUG nova.compute.manager [req-37b16736-5f7e-4112-ba98-9bc3913dd257 req-a439f0eb-af24-4af4-bac8-cf9465c95e43 service nova] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Received event network-vif-deleted-9c5aba0e-5fe6-41fd-a20c-fa9523d5a86f {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 671.897799] env[66641]: DEBUG nova.policy [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7279b628bf1646d8bd27416d1f52cfdb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196b2ed18f7142a18f9093869dadb46e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 672.021692] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.022192] env[66641]: WARNING openstack [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.061583] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145674, 'name': PowerOffVM_Task, 'duration_secs': 0.220185} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.061583] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 672.061737] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 17 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 672.069996] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baef0a39-ed9d-484d-9173-d0a28cd80752 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.079821] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b58478-ef94-4fc6-9de7-3dd19a55a6d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.121342] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecceb3ae-bfb3-4eca-a3dc-8f489c574c97 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.138769] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145675, 'name': Rename_Task, 'duration_secs': 0.208146} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.138769] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 672.139025] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9082e049-b9f7-4ce7-8446-49976e540e23 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.144070] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4f6c449b-29af-429b-9e9b-62d5d0df1bc7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.159053] env[66641]: DEBUG nova.compute.provider_tree [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.165606] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145676, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.165860] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 672.165860] env[66641]: value = "task-5145677" [ 672.165860] env[66641]: _type = "Task" [ 672.165860] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.180112] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145677, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.330904] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.331053] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 672.331269] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 672.453428] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Successfully created port: f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 672.570901] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 672.571204] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 672.571433] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 672.571718] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 672.571939] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 672.572327] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 672.572412] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.572621] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 672.572817] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 672.573052] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 672.573275] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 672.581873] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd77e901-6f48-470e-a2c0-b617c3a8045b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.606770] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 672.606770] env[66641]: value = "task-5145678" [ 672.606770] env[66641]: _type = "Task" [ 672.606770] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.618499] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145678, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.650703] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145676, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.670202] env[66641]: DEBUG nova.scheduler.client.report [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 672.684648] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145677, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.792675] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 672.828381] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 672.828381] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 672.828381] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 672.828565] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 672.828565] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 672.828565] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 672.828666] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 672.829264] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 672.829264] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 672.829542] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 672.829954] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 672.831317] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c205b2d-32db-4449-943a-e15eca9c39c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.839585] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 672.840020] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 672.845547] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 672.848996] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e63a9f-1ad1-4dcf-889b-c0791addfef0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.962985] env[66641]: DEBUG nova.network.neutron [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Updated VIF entry in instance network info cache for port d3f790d9-7633-46ba-842e-725e9680d0ea. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 672.963660] env[66641]: DEBUG nova.network.neutron [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Updating instance_info_cache with network_info: [{"id": "d3f790d9-7633-46ba-842e-725e9680d0ea", "address": "fa:16:3e:a7:27:b6", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f790d9-76", "ovs_interfaceid": "d3f790d9-7633-46ba-842e-725e9680d0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.121927] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145678, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.147032] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145676, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.153144] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.153588] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.182079] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.420s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 673.182737] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 673.185982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.910s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 673.188030] env[66641]: INFO nova.compute.claims [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 673.200313] env[66641]: DEBUG oslo_vmware.api [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145677, 'name': PowerOnVM_Task, 'duration_secs': 0.535137} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.200440] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 673.200645] env[66641]: INFO nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Took 11.40 seconds to spawn the instance on the hypervisor. [ 673.200846] env[66641]: DEBUG nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 673.201788] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-639bc0da-1dc4-4954-a249-4d6d8cd57bd1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.232279] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.232279] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.417304] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Updating instance_info_cache with network_info: [{"id": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "address": "fa:16:3e:3f:01:c2", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fd71b07-2d", "ovs_interfaceid": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 673.470274] env[66641]: DEBUG oslo_concurrency.lockutils [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] Releasing lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.470274] env[66641]: DEBUG nova.compute.manager [req-45d1b6b1-5ba4-49a7-8b78-345dff1c375d req-6b1cd8e4-310d-4795-b6d1-f61e2fe510cb service nova] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Received event network-vif-deleted-46c2027b-2796-42b8-9167-719bceab5640 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 673.621792] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145678, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.650022] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145676, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.85055} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.650022] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 673.650022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 673.650022] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f40d72d8-d7ab-4507-bbcb-0c20cbae675c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.657770] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 673.657770] env[66641]: value = "task-5145679" [ 673.657770] env[66641]: _type = "Task" [ 673.657770] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.672582] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.698765] env[66641]: DEBUG nova.compute.utils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 673.701549] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 673.701926] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 673.702405] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.702697] env[66641]: WARNING neutronclient.v2_0.client [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 673.703971] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 673.704126] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 673.723795] env[66641]: INFO nova.compute.manager [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Took 20.41 seconds to build instance. [ 673.804460] env[66641]: DEBUG nova.policy [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7279b628bf1646d8bd27416d1f52cfdb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196b2ed18f7142a18f9093869dadb46e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 673.922517] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 673.922517] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Instance network_info: |[{"id": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "address": "fa:16:3e:3f:01:c2", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fd71b07-2d", "ovs_interfaceid": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 673.922995] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:01:c2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2fd71b07-2d9b-426a-a09a-715cea2538e3', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 673.930534] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Creating folder: Project (196b2ed18f7142a18f9093869dadb46e). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 673.930864] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b7bdb39c-04f9-4502-aac2-23ed677accb7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.943392] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Created folder: Project (196b2ed18f7142a18f9093869dadb46e) in parent group-v1000566. [ 673.943392] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Creating folder: Instances. Parent ref: group-v1000611. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 673.943392] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21d36983-6a98-448a-bf73-de446deb6bec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.958027] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Created folder: Instances in parent group-v1000611. [ 673.958821] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 673.958821] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 673.958980] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc4be4c7-9b6f-4284-bb74-7b003e5b886d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.986816] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 673.986816] env[66641]: value = "task-5145682" [ 673.986816] env[66641]: _type = "Task" [ 673.986816] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.997390] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145682, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.118578] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145678, 'name': ReconfigVM_Task, 'duration_secs': 1.327448} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.119077] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 33 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 674.168848] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076573} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.169894] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 674.170761] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f762b4bf-1daa-49e3-996e-549cad89d0bb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.192538] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfiguring VM instance instance-0000000b to attach disk [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 674.195659] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Successfully updated port: f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 674.197283] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89167b25-580c-41f5-9b15-5a18d5284f2d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.212571] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.212762] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.212931] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 674.215165] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 674.225746] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f90efe99-f7a7-4bdb-bd9c-62076a710e5c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.928s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 674.229231] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 674.229231] env[66641]: value = "task-5145683" [ 674.229231] env[66641]: _type = "Task" [ 674.229231] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.241614] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145683, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.446758] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Successfully created port: dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 674.453517] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff83d49-b4c1-4a56-898f-1998ba31fc66 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.462670] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cede7527-c1b4-4454-abe9-8bc0308a5d7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.500265] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e478dea7-2010-4d28-8a2c-a9d239809c21 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.513662] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6bf213-c5cb-44fa-a0d3-c75e1553592f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.517787] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145682, 'name': CreateVM_Task, 'duration_secs': 0.453596} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.517992] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 674.521070] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.521070] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.526676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.526676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 674.526952] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 674.536244] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c842c97a-2c37-47e2-a7d9-556b8dd32976 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.538293] env[66641]: DEBUG nova.compute.provider_tree [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.543664] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 674.543664] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24460-d5b5-4558-2b6d-47513cb8657a" [ 674.543664] env[66641]: _type = "Task" [ 674.543664] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.553376] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24460-d5b5-4558-2b6d-47513cb8657a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.626104] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 674.626489] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 674.626671] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 674.626851] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 674.630381] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 674.630381] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 674.630381] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 674.630381] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 674.630381] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 674.630614] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 674.630614] env[66641]: DEBUG nova.virt.hardware [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 674.634699] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfiguring VM instance instance-00000006 to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 674.634994] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5867c41c-7add-4e12-bb8e-efddae95eeee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.654629] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 674.654629] env[66641]: value = "task-5145684" [ 674.654629] env[66641]: _type = "Task" [ 674.654629] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.666352] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145684, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.726690] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 674.726690] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 674.733388] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 674.753159] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145683, 'name': ReconfigVM_Task, 'duration_secs': 0.399513} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.753446] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfigured VM instance instance-0000000b to attach disk [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 674.754184] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1ab5757-f634-4528-ace7-b68aecb88ad1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.763480] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 674.763480] env[66641]: value = "task-5145685" [ 674.763480] env[66641]: _type = "Task" [ 674.763480] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.774715] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145685, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.044037] env[66641]: DEBUG nova.scheduler.client.report [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 675.057340] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24460-d5b5-4558-2b6d-47513cb8657a, 'name': SearchDatastore_Task, 'duration_secs': 0.015101} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.058208] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 675.060625] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 675.060625] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.060625] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.060625] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 675.060625] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e4e5b4ec-a80f-4627-9db7-2069bb011c0e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.071335] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 675.071335] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 675.071335] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f28d64e1-4c06-4902-8e64-4b3249f4440e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.078198] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 675.078198] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524b7b0f-003e-69ef-c7eb-f505d1c7c011" [ 675.078198] env[66641]: _type = "Task" [ 675.078198] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.087818] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524b7b0f-003e-69ef-c7eb-f505d1c7c011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.166762] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145684, 'name': ReconfigVM_Task, 'duration_secs': 0.227076} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.167083] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfigured VM instance instance-00000006 to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 675.168085] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38892636-b4ac-45f6-a2ea-4b94da07545c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.199677] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 675.200494] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc6539e5-2b49-49d1-8990-d87814f03dd1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.228188] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 675.228188] env[66641]: value = "task-5145686" [ 675.228188] env[66641]: _type = "Task" [ 675.228188] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.237324] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 675.243539] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145686, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.268577] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.268972] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.278418] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 675.278632] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 675.278777] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 675.278946] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 675.279098] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 675.279240] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 675.279446] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 675.279624] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 675.279736] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 675.279891] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 675.280133] env[66641]: DEBUG nova.virt.hardware [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 675.281712] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df5950b-7887-42c4-a6a8-4f8b09a93205 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.287160] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145685, 'name': Rename_Task, 'duration_secs': 0.191508} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.287797] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 675.288105] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-14280f57-aab9-4537-a489-54b573514b3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.297123] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be5bc2f-a9e4-48fe-90f3-1f1d60c49bb8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.302919] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 675.302919] env[66641]: value = "task-5145687" [ 675.302919] env[66641]: _type = "Task" [ 675.302919] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.339786] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145687, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.461195] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.461195] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.552128] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.366s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.553063] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 675.557369] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.758s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.558133] env[66641]: DEBUG nova.objects.instance [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lazy-loading 'resources' on Instance uuid 47aaed25-542b-4ceb-9adf-6a4953c8c95d {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 675.591713] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524b7b0f-003e-69ef-c7eb-f505d1c7c011, 'name': SearchDatastore_Task, 'duration_secs': 0.011199} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.592701] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a46bfb4-31cc-4ade-b61c-3cb49c983706 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.600883] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 675.600883] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5234bf64-1ae5-02d6-1f47-2a8114a7a54a" [ 675.600883] env[66641]: _type = "Task" [ 675.600883] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.611172] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5234bf64-1ae5-02d6-1f47-2a8114a7a54a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.740630] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145686, 'name': ReconfigVM_Task, 'duration_secs': 0.396954} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.740747] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Reconfigured VM instance instance-00000006 to attach disk [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949/dfa8c73b-db57-42a9-a9a4-cf812f5b2949.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 675.740980] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 50 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 675.785861] env[66641]: DEBUG nova.compute.manager [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Received event network-vif-plugged-4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 675.786100] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Acquiring lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 675.786304] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 675.787212] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 675.788229] env[66641]: DEBUG nova.compute.manager [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] No waiting events found dispatching network-vif-plugged-4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 675.788767] env[66641]: WARNING nova.compute.manager [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Received unexpected event network-vif-plugged-4c64a4dc-a032-4361-b383-ea4b1d7c3728 for instance with vm_state active and task_state None. [ 675.789232] env[66641]: DEBUG nova.compute.manager [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Received event network-changed-4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 675.789460] env[66641]: DEBUG nova.compute.manager [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Refreshing instance network info cache due to event network-changed-4c64a4dc-a032-4361-b383-ea4b1d7c3728. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 675.789841] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Acquiring lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.790377] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Acquired lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 675.790542] env[66641]: DEBUG nova.network.neutron [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Refreshing network info cache for port 4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 675.799423] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 675.799423] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 675.822134] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145687, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.921657] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Updating instance_info_cache with network_info: [{"id": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "address": "fa:16:3e:26:f9:99", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3fb5c31-b4", "ovs_interfaceid": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 675.964778] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 676.061642] env[66641]: DEBUG nova.compute.utils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 676.063296] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 676.063787] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 676.064415] env[66641]: WARNING neutronclient.v2_0.client [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.065288] env[66641]: WARNING neutronclient.v2_0.client [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.066466] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.066553] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.104935] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Successfully updated port: dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 676.120211] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5234bf64-1ae5-02d6-1f47-2a8114a7a54a, 'name': SearchDatastore_Task, 'duration_secs': 0.016835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.120211] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.120342] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 434e66b1-5ea7-43d2-9b41-bdeda2a312d7/434e66b1-5ea7-43d2-9b41-bdeda2a312d7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 676.120703] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-932f7d5a-b7f0-4560-9981-5de2377d50e3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.128793] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 676.128793] env[66641]: value = "task-5145688" [ 676.128793] env[66641]: _type = "Task" [ 676.128793] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.142349] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145688, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.251162] env[66641]: DEBUG nova.policy [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d0eeef5974e470fa2a3b78e7b25bf9a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '82383b2f8cf2490ebdf92495d9dc20e6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 676.258824] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7c5d95b-f2c5-4bf5-9c81-2aad2f42ab8f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.280259] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc1ecc3-23ff-4ea7-aca2-9b414e4f8c35 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.302243] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 67 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 676.309468] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.309843] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.327949] env[66641]: DEBUG oslo_vmware.api [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145687, 'name': PowerOnVM_Task, 'duration_secs': 0.774807} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.329512] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 676.329512] env[66641]: DEBUG nova.compute.manager [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 676.329512] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d51e00e-3a52-4435-9dbc-25f0a0b6309c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.425926] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 676.426261] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Instance network_info: |[{"id": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "address": "fa:16:3e:26:f9:99", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3fb5c31-b4", "ovs_interfaceid": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 676.427581] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:f9:99', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 676.435139] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 676.439013] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 676.440344] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-09b3dbf8-4c10-4a03-afed-a01981aaa18a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.462662] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 676.462662] env[66641]: value = "task-5145689" [ 676.462662] env[66641]: _type = "Task" [ 676.462662] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.476216] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145689, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.485219] env[66641]: DEBUG nova.compute.manager [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Received event network-changed {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 676.487016] env[66641]: DEBUG nova.compute.manager [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Refreshing instance network info cache due to event network-changed. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 676.487016] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Acquiring lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.487016] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Acquired lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.487016] env[66641]: DEBUG nova.network.neutron [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 676.502675] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.579736] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 676.585055] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c4a383-2f06-4ef1-945f-d6d2e7d12dbe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.593325] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d91b9b-51b4-436e-86e8-5c484ec7489b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.634025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.634025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.634025] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 676.637496] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db498120-8976-4d4c-b29e-f1028a745357 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.647802] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145688, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.653336] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e959e08-89a0-49ba-8a10-c7bafd4c2559 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.672495] env[66641]: DEBUG nova.compute.provider_tree [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.748924] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Successfully created port: f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 676.777079] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.777588] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.819777] env[66641]: WARNING neutronclient.v2_0.client [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 676.850587] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 676.974281] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145689, 'name': CreateVM_Task, 'duration_secs': 0.46511} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.974485] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 676.975230] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.975602] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 676.981421] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.981649] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 676.981993] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 676.982513] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-adb579ed-214d-4fe0-a2e5-f060f6c39424 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.990719] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 676.991249] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.001508] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 677.001508] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5210c2b9-a21b-4eae-4e37-2fbc0031c30c" [ 677.001508] env[66641]: _type = "Task" [ 677.001508] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.014397] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5210c2b9-a21b-4eae-4e37-2fbc0031c30c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.077786] env[66641]: DEBUG nova.network.neutron [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Port 5ebec271-9d57-41f2-ba8c-59edb82c50db binding to destination host cpu-1 is already ACTIVE {{(pid=66641) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 677.142768] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.143174] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.148135] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 677.156763] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145688, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655833} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.157014] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 434e66b1-5ea7-43d2-9b41-bdeda2a312d7/434e66b1-5ea7-43d2-9b41-bdeda2a312d7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 677.157233] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.157477] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8825b33c-f635-433b-bbbf-c9b1264b169d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.165182] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 677.165182] env[66641]: value = "task-5145690" [ 677.165182] env[66641]: _type = "Task" [ 677.165182] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.174871] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145690, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.175886] env[66641]: DEBUG nova.scheduler.client.report [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 677.380422] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 677.380843] env[66641]: WARNING openstack [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 677.513319] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5210c2b9-a21b-4eae-4e37-2fbc0031c30c, 'name': SearchDatastore_Task, 'duration_secs': 0.038847} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.513623] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 677.514389] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 677.514693] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.514879] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 677.515101] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 677.515672] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6617f6b2-d22e-46d5-8c3f-b81bdcf0e2e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.530111] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 677.530355] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 677.531539] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5e16a64-e719-4aab-b686-f0c47232946c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.539210] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 677.539210] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e75a2c-fefd-e13e-c4c7-b798c39b0f41" [ 677.539210] env[66641]: _type = "Task" [ 677.539210] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.551852] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e75a2c-fefd-e13e-c4c7-b798c39b0f41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.592709] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 677.633632] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 677.633883] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 677.634038] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 677.634219] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 677.636183] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 677.636183] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 677.636183] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.636183] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 677.636183] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 677.636555] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 677.636555] env[66641]: DEBUG nova.virt.hardware [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 677.636555] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf700bf-2fce-43eb-8a24-a20327a8ef04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.645232] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5b197a3-0b99-4e5e-b9ed-8f2751bd3c91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.676166] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145690, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069077} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.676425] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 677.677121] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc286c16-ed30-42b0-a8a6-8566698e6a37 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.680681] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.124s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 677.683975] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.171s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 677.683975] env[66641]: DEBUG nova.objects.instance [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lazy-loading 'resources' on Instance uuid 6b19204d-032c-410b-bb84-1af23d142edc {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 677.704868] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Reconfiguring VM instance instance-0000000f to attach disk [datastore2] 434e66b1-5ea7-43d2-9b41-bdeda2a312d7/434e66b1-5ea7-43d2-9b41-bdeda2a312d7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 677.705954] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3a41dc08-3423-4856-adc6-9451694ffab8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.721458] env[66641]: INFO nova.scheduler.client.report [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Deleted allocations for instance 47aaed25-542b-4ceb-9adf-6a4953c8c95d [ 677.732415] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 677.732415] env[66641]: value = "task-5145691" [ 677.732415] env[66641]: _type = "Task" [ 677.732415] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.744929] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145691, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.051799] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e75a2c-fefd-e13e-c4c7-b798c39b0f41, 'name': SearchDatastore_Task, 'duration_secs': 0.021743} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.052609] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-707efc38-a8fd-4daf-831d-b37b1b6a8f0c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.058270] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 678.058270] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526495b9-c42c-dc52-3012-eadeca466232" [ 678.058270] env[66641]: _type = "Task" [ 678.058270] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.066861] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526495b9-c42c-dc52-3012-eadeca466232, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.099941] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.100262] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.100478] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.230158] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0ea1738d-9f86-42b8-b793-35054401393d tempest-VolumesAdminNegativeTest-1230207175 tempest-VolumesAdminNegativeTest-1230207175-project-member] Lock "47aaed25-542b-4ceb-9adf-6a4953c8c95d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.162s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.243625] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145691, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.309752] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Successfully updated port: f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 678.402678] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f7bb18-9b8b-4905-b39c-fd3f19655900 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.411741] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078ad87f-58a8-4e02-a3ac-be132b6e97b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.418081] env[66641]: DEBUG nova.compute.manager [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Received event network-vif-plugged-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 678.418320] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] Acquiring lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 678.418522] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 678.418681] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 678.419029] env[66641]: DEBUG nova.compute.manager [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] No waiting events found dispatching network-vif-plugged-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 678.419029] env[66641]: WARNING nova.compute.manager [req-bd65f8c1-7205-451d-b87a-87d5da83a4b6 req-34c8e719-b1ce-448b-8d70-594c6b0f6b00 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Received unexpected event network-vif-plugged-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 for instance with vm_state building and task_state spawning. [ 678.445530] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443befcd-2e83-45a1-8ac7-bffb1c6602ee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.454956] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22808e1-4cd9-4aba-931a-4f27858bafb5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.473296] env[66641]: DEBUG nova.compute.provider_tree [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.568838] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526495b9-c42c-dc52-3012-eadeca466232, 'name': SearchDatastore_Task, 'duration_secs': 0.01828} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.569144] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 678.569407] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d14528cb-f26e-4fe5-8bf9-34900571fe03/d14528cb-f26e-4fe5-8bf9-34900571fe03.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 678.569672] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-20253204-32bf-437a-ac6c-206926d52bc9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.581667] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 678.581667] env[66641]: value = "task-5145692" [ 678.581667] env[66641]: _type = "Task" [ 678.581667] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.591674] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145692, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.593422] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.593531] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.605130] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 678.605497] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 678.744859] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145691, 'name': ReconfigVM_Task, 'duration_secs': 0.859054} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.745313] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Reconfigured VM instance instance-0000000f to attach disk [datastore2] 434e66b1-5ea7-43d2-9b41-bdeda2a312d7/434e66b1-5ea7-43d2-9b41-bdeda2a312d7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.746322] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b3011c2f-9aed-4543-8e3b-8081f2ccb14f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.754732] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 678.754732] env[66641]: value = "task-5145693" [ 678.754732] env[66641]: _type = "Task" [ 678.754732] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.765033] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145693, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.813982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.814179] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 678.814359] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 678.980294] env[66641]: DEBUG nova.scheduler.client.report [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 679.095486] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145692, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.115348] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.115830] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.270030] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145693, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.318593] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.318865] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.325736] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 679.356167] env[66641]: DEBUG nova.network.neutron [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Updated VIF entry in instance network info cache for port 4c64a4dc-a032-4361-b383-ea4b1d7c3728. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 679.356543] env[66641]: DEBUG nova.network.neutron [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Updating instance_info_cache with network_info: [{"id": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "address": "fa:16:3e:f8:4e:ac", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4c64a4dc-a0", "ovs_interfaceid": "4c64a4dc-a032-4361-b383-ea4b1d7c3728", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.472837] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.473216] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 679.473216] env[66641]: DEBUG nova.network.neutron [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 679.486255] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 679.489060] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.714s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 679.491228] env[66641]: INFO nova.compute.claims [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.524988] env[66641]: INFO nova.scheduler.client.report [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Deleted allocations for instance 6b19204d-032c-410b-bb84-1af23d142edc [ 679.548618] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.549019] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.582772] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.583211] env[66641]: WARNING openstack [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.602540] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145692, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.587512} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.602540] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d14528cb-f26e-4fe5-8bf9-34900571fe03/d14528cb-f26e-4fe5-8bf9-34900571fe03.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 679.602540] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 679.602540] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c5026e6-3b26-42c5-a98b-d50c6c1ed489 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.611260] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 679.611260] env[66641]: value = "task-5145694" [ 679.611260] env[66641]: _type = "Task" [ 679.611260] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.629038] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145694, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.681803] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.682226] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.771870] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145693, 'name': Rename_Task, 'duration_secs': 0.604239} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.772183] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 679.772435] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-767daaf0-949a-456d-b4e9-3d39fb20550f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.780079] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 679.780079] env[66641]: value = "task-5145695" [ 679.780079] env[66641]: _type = "Task" [ 679.780079] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.789292] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.847390] env[66641]: DEBUG nova.network.neutron [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Updating instance_info_cache with network_info: [{"id": "dbb7738c-a759-4460-8f36-9f673e12b204", "address": "fa:16:3e:8d:c8:a4", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbb7738c-a7", "ovs_interfaceid": "dbb7738c-a759-4460-8f36-9f673e12b204", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.859718] env[66641]: DEBUG oslo_concurrency.lockutils [req-85e81684-480f-4555-9b41-8763f231b96c req-aafb6991-8f46-4640-8e91-762866df6dfc service nova] Releasing lock "refresh_cache-695d1824-cf34-46c1-b623-b474d2f1f8ef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 679.924776] env[66641]: DEBUG nova.network.neutron [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Updating instance_info_cache with network_info: [{"id": "d3f790d9-7633-46ba-842e-725e9680d0ea", "address": "fa:16:3e:a7:27:b6", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.228", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3f790d9-76", "ovs_interfaceid": "d3f790d9-7633-46ba-842e-725e9680d0ea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 679.978535] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.978923] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 679.990069] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 679.991399] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.049330] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e700f297-e9b9-4b01-a710-2439689c4282 tempest-ListServerFiltersTestJSON-337841595 tempest-ListServerFiltersTestJSON-337841595-project-member] Lock "6b19204d-032c-410b-bb84-1af23d142edc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.771s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 680.126824] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145694, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.121883} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.126824] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 680.126824] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a767c0d7-50d4-4d7e-a142-ec8e086e82ad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.157075] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Reconfiguring VM instance instance-00000010 to attach disk [datastore2] d14528cb-f26e-4fe5-8bf9-34900571fe03/d14528cb-f26e-4fe5-8bf9-34900571fe03.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 680.158641] env[66641]: DEBUG nova.network.neutron [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Updating instance_info_cache with network_info: [{"id": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "address": "fa:16:3e:84:b1:ad", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4cb47fb-c2", "ovs_interfaceid": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.159921] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98eaf1bc-6969-49bc-a3dc-b394a7dcc9d4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.176337] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.176794] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.185012] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.185396] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Instance network_info: |[{"id": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "address": "fa:16:3e:84:b1:ad", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4cb47fb-c2", "ovs_interfaceid": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 680.186246] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:84:b1:ad', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.194450] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 680.195801] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 680.196116] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b152d2f-5a3e-4eea-9d52-3475f9873ea4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.218327] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 680.218327] env[66641]: value = "task-5145696" [ 680.218327] env[66641]: _type = "Task" [ 680.218327] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.228032] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.228032] env[66641]: value = "task-5145697" [ 680.228032] env[66641]: _type = "Task" [ 680.228032] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.233151] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145696, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.238719] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145697, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.295390] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.305616] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.306031] env[66641]: WARNING openstack [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.351954] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.351954] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Instance network_info: |[{"id": "dbb7738c-a759-4460-8f36-9f673e12b204", "address": "fa:16:3e:8d:c8:a4", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbb7738c-a7", "ovs_interfaceid": "dbb7738c-a759-4460-8f36-9f673e12b204", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 680.354853] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:c8:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbb7738c-a759-4460-8f36-9f673e12b204', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.359134] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 680.359384] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 680.359610] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0978d37-4f83-4069-9aae-23a4b605f12f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.378440] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "a2d567c9-09af-44e1-aad2-c8f394a09714" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.378597] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.387036] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.387036] env[66641]: value = "task-5145698" [ 680.387036] env[66641]: _type = "Task" [ 680.387036] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.396159] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145698, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.424720] env[66641]: DEBUG oslo_concurrency.lockutils [None req-8cc8a702-8bd2-408e-b5f8-c5603c0139ea tempest-ServerExternalEventsTest-456291400 tempest-ServerExternalEventsTest-456291400-project] Releasing lock "refresh_cache-e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.710813] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 680.711292] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 680.739435] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145696, 'name': ReconfigVM_Task, 'duration_secs': 0.417772} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.739743] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Reconfigured VM instance instance-00000010 to attach disk [datastore2] d14528cb-f26e-4fe5-8bf9-34900571fe03/d14528cb-f26e-4fe5-8bf9-34900571fe03.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.740461] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-228a5ade-c71c-4168-b4c4-4d9fe5ca5fa9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.746527] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145697, 'name': CreateVM_Task, 'duration_secs': 0.389191} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.746729] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 680.750577] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.751055] env[66641]: WARNING openstack [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.757267] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.757267] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.757626] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 680.760213] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b8e718c-a925-4dbc-99f7-0b643782d4f0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.762313] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 680.762313] env[66641]: value = "task-5145699" [ 680.762313] env[66641]: _type = "Task" [ 680.762313] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.768819] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 680.768819] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ea7287-80c2-3c08-f1ee-194284feb81c" [ 680.768819] env[66641]: _type = "Task" [ 680.768819] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.780530] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145699, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.792609] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ea7287-80c2-3c08-f1ee-194284feb81c, 'name': SearchDatastore_Task, 'duration_secs': 0.018018} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.794468] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 680.794468] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 680.794468] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.794468] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.794763] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 680.794763] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-28c51a73-9c65-480f-ab41-9577da3f57cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.803526] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.814631] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 680.814932] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 680.816165] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-629fde6e-645c-4195-b53e-75f367b505fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.828784] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 680.828784] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f32f67-b0d9-83c7-1407-3e2eabd1e45d" [ 680.828784] env[66641]: _type = "Task" [ 680.828784] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.839412] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f32f67-b0d9-83c7-1407-3e2eabd1e45d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.851885] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f6371c-f2e8-41d2-94e5-d8c78010bd17 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.860232] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b150300-3063-43e4-8bf1-cc27fd229773 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.893556] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 680.900425] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-342085f8-6121-4351-96c0-72e52807ab04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.913307] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9cb36f-707c-4cd8-a5d1-4ecd5f52c83a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.919021] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145698, 'name': CreateVM_Task, 'duration_secs': 0.397171} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.919239] env[66641]: DEBUG nova.network.neutron [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 680.920468] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 680.923351] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 680.923351] env[66641]: WARNING openstack [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 680.927533] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.927695] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 680.928762] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 680.938195] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-faad6a27-50f4-44eb-a2b8-4adc88f28859 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.940378] env[66641]: DEBUG nova.compute.provider_tree [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.946476] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 680.946476] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5296721d-0df8-7984-e2c1-a710b93a006e" [ 680.946476] env[66641]: _type = "Task" [ 680.946476] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.956076] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5296721d-0df8-7984-e2c1-a710b93a006e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.214609] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 681.274707] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145699, 'name': Rename_Task, 'duration_secs': 0.155772} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.278674] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 681.278674] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-26663283-9343-436c-8c49-e753bed227e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.286860] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 681.286860] env[66641]: value = "task-5145700" [ 681.286860] env[66641]: _type = "Task" [ 681.286860] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.303376] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.310857] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145700, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.340427] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f32f67-b0d9-83c7-1407-3e2eabd1e45d, 'name': SearchDatastore_Task, 'duration_secs': 0.02528} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.341330] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a89c097d-027e-4387-985b-1340c5d6ada1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.348478] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 681.348478] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258b4cd-c156-866c-2823-c79fe9402f4f" [ 681.348478] env[66641]: _type = "Task" [ 681.348478] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.357362] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258b4cd-c156-866c-2823-c79fe9402f4f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.422998] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.428925] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.444285] env[66641]: DEBUG nova.scheduler.client.report [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 681.463368] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5296721d-0df8-7984-e2c1-a710b93a006e, 'name': SearchDatastore_Task, 'duration_secs': 0.037811} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.463368] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.463368] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 681.463368] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.754702] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 681.807365] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.807365] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145700, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.872703] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258b4cd-c156-866c-2823-c79fe9402f4f, 'name': SearchDatastore_Task, 'duration_secs': 0.036584} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.874624] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 681.874919] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 22899269-28fc-446e-9533-1f6862a9e8db/22899269-28fc-446e-9533-1f6862a9e8db.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 681.875930] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 681.876196] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 681.876456] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-289556dc-12f6-4150-b5b0-be99e7a4f60a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.879115] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-70fdf54d-878c-4552-bb2c-b5ed20fc2ecd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.888098] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 681.888098] env[66641]: value = "task-5145701" [ 681.888098] env[66641]: _type = "Task" [ 681.888098] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.895283] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 681.895283] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 681.895686] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8f67938-772f-4920-907a-f67aa061216c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.901749] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145701, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.905816] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 681.905816] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24dd2-3162-3bd2-26d2-40d178a37d28" [ 681.905816] env[66641]: _type = "Task" [ 681.905816] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.915251] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24dd2-3162-3bd2-26d2-40d178a37d28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.952635] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48bf8425-32f5-4298-ae5c-335aee59feaa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.956795] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 681.958631] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 681.962084] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.459s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 681.963648] env[66641]: INFO nova.compute.claims [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 681.986485] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b898f657-025b-41eb-8dc7-7d112d93a56b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.995756] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 83 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 682.303451] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145695, 'name': PowerOnVM_Task, 'duration_secs': 2.099821} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.304337] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.304455] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Took 11.98 seconds to spawn the instance on the hypervisor. [ 682.304668] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 682.305953] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ba8baa-42ca-431f-a20d-97a34cec0422 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.311958] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145700, 'name': PowerOnVM_Task, 'duration_secs': 0.705756} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.313157] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 682.313157] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Took 9.52 seconds to spawn the instance on the hypervisor. [ 682.313157] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 682.313829] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf91f81-d5b7-4de8-ada2-4f41c3ff545d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.376910] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.376910] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.376910] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 682.376910] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 682.377254] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 682.379733] env[66641]: INFO nova.compute.manager [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Terminating instance [ 682.401603] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145701, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.420412] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e24dd2-3162-3bd2-26d2-40d178a37d28, 'name': SearchDatastore_Task, 'duration_secs': 0.019674} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.420665] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dddef0d8-5eb8-485e-8b9b-18e143e399b2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.430200] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 682.430200] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525b13ab-2ead-0c18-49a7-3874dff366e0" [ 682.430200] env[66641]: _type = "Task" [ 682.430200] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.441140] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525b13ab-2ead-0c18-49a7-3874dff366e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.472547] env[66641]: DEBUG nova.compute.utils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 682.473039] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 682.473322] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 682.473550] env[66641]: WARNING neutronclient.v2_0.client [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.473826] env[66641]: WARNING neutronclient.v2_0.client [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 682.477020] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 682.477020] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 682.504461] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 682.505181] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10ac17e8-ae3e-47bb-827e-f8d4db81d6c2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.515630] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 682.515630] env[66641]: value = "task-5145702" [ 682.515630] env[66641]: _type = "Task" [ 682.515630] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.527306] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145702, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.843708] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Took 26.87 seconds to build instance. [ 682.845622] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Took 26.92 seconds to build instance. [ 682.884833] env[66641]: DEBUG nova.compute.manager [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 682.884833] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 682.885628] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b16d82-cbd1-41e7-9fb5-a1abab711783 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.896369] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 682.896991] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-73adc88c-9377-4be6-adf9-5a6027bfbaca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.903020] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145701, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.860032} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.903651] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 22899269-28fc-446e-9533-1f6862a9e8db/22899269-28fc-446e-9533-1f6862a9e8db.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 682.903864] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 682.904127] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-645be821-6942-4a63-b17c-3d58b9454ea3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.909074] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 682.909074] env[66641]: value = "task-5145703" [ 682.909074] env[66641]: _type = "Task" [ 682.909074] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.915194] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 682.915194] env[66641]: value = "task-5145704" [ 682.915194] env[66641]: _type = "Task" [ 682.915194] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.919147] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145703, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.932028] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145704, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.946078] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525b13ab-2ead-0c18-49a7-3874dff366e0, 'name': SearchDatastore_Task, 'duration_secs': 0.065357} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.946078] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 682.946078] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] fb1bf80c-b243-4c25-9aaf-c459ca926090/fb1bf80c-b243-4c25-9aaf-c459ca926090.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 682.946078] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d4dac56-ced0-41a3-ad40-58efea950ed1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.954893] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 682.954893] env[66641]: value = "task-5145705" [ 682.954893] env[66641]: _type = "Task" [ 682.954893] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.967610] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.983874] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 683.039128] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145702, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.186754] env[66641]: DEBUG nova.policy [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d6c5fbc65414faebe5c3c9bb4ec4309', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ab82c335b15478788a818319bf11c03', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 683.257777] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7533bbc9-fb40-43bb-9772-16c226820b33 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.266258] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40072777-887b-44b5-ad30-c49e5cfe9563 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.309146] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a523ac15-d820-401c-aa74-b1982864cbc3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.314480] env[66641]: DEBUG nova.compute.manager [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Received event network-vif-plugged-2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 683.314664] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Acquiring lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 683.314876] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 683.315058] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.315224] env[66641]: DEBUG nova.compute.manager [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] No waiting events found dispatching network-vif-plugged-2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 683.315429] env[66641]: WARNING nova.compute.manager [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Received unexpected event network-vif-plugged-2fd71b07-2d9b-426a-a09a-715cea2538e3 for instance with vm_state active and task_state None. [ 683.315572] env[66641]: DEBUG nova.compute.manager [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Received event network-changed-2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 683.315715] env[66641]: DEBUG nova.compute.manager [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Refreshing instance network info cache due to event network-changed-2fd71b07-2d9b-426a-a09a-715cea2538e3. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 683.315940] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Acquiring lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.316082] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Acquired lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 683.316296] env[66641]: DEBUG nova.network.neutron [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Refreshing network info cache for port 2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 683.325628] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee40173f-c7bb-4846-93bd-e3c68bfd9130 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.346990] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.389s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.347728] env[66641]: DEBUG nova.compute.provider_tree [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.352596] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.444s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 683.423302] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145703, 'name': PowerOffVM_Task, 'duration_secs': 0.383693} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.427030] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 683.427176] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 683.427923] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50b9e780-b048-452f-a45e-3677073b4795 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.437534] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145704, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082102} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.437792] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 683.440253] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af37ab0-a7af-4398-a6ca-a7e29dc671d2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.467782] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Reconfiguring VM instance instance-00000012 to attach disk [datastore2] 22899269-28fc-446e-9533-1f6862a9e8db/22899269-28fc-446e-9533-1f6862a9e8db.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 683.472126] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d86b8932-85b9-4053-bb05-a9eeab3d9d3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.499524] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145705, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.500652] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 683.500652] env[66641]: value = "task-5145707" [ 683.500652] env[66641]: _type = "Task" [ 683.500652] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.511605] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145707, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.523014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 683.523014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 683.523446] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Deleting the datastore file [datastore1] e7bfb26e-b326-4a39-8f20-79dbbfb74eb8 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 683.524057] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c52f49dc-5f33-47a8-b83e-0ea84e3cd378 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.531327] env[66641]: DEBUG oslo_vmware.api [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145702, 'name': PowerOnVM_Task, 'duration_secs': 0.580025} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.531663] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 683.531844] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7f42b712-dc0f-45cb-8831-734899ce19af tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance 'dfa8c73b-db57-42a9-a9a4-cf812f5b2949' progress to 100 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 683.540437] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for the task: (returnval){ [ 683.540437] env[66641]: value = "task-5145708" [ 683.540437] env[66641]: _type = "Task" [ 683.540437] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.557420] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.819831] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 683.820436] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 683.851421] env[66641]: DEBUG nova.scheduler.client.report [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 683.927818] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Successfully created port: 653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 683.969236] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.698253} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.969609] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] fb1bf80c-b243-4c25-9aaf-c459ca926090/fb1bf80c-b243-4c25-9aaf-c459ca926090.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 683.969726] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 683.969955] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-13612e68-3b63-4a97-ae88-166e8ed6ba11 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.978913] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 683.978913] env[66641]: value = "task-5145709" [ 683.978913] env[66641]: _type = "Task" [ 683.978913] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.989716] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.005696] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 684.016072] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145707, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.018811] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.019732] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.054799] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145708, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.075417] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 684.075714] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 684.075908] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 684.076125] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 684.076287] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 684.076434] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 684.076639] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.076786] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 684.076943] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 684.078054] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 684.078054] env[66641]: DEBUG nova.virt.hardware [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 684.079432] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cd56f0-acbc-4657-8c50-ecdc563361bd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.087857] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24562a6f-95fa-4330-833a-e1170a705f14 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.246454] env[66641]: INFO nova.compute.manager [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Rebuilding instance [ 684.314018] env[66641]: DEBUG nova.compute.manager [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 684.314917] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077c95bd-ed40-4964-aace-34096833e974 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.361032] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.398s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.361032] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 684.363288] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 7.513s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.363872] env[66641]: DEBUG nova.objects.instance [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 684.492031] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.162713} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.492031] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.495277] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb05710f-d8b7-4deb-8540-b1b0faa37b07 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.522342] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Reconfiguring VM instance instance-00000011 to attach disk [datastore2] fb1bf80c-b243-4c25-9aaf-c459ca926090/fb1bf80c-b243-4c25-9aaf-c459ca926090.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.530065] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-07a45fb0-2304-42b4-b88d-f1f4d8526a29 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.551316] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.556716] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.561470] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.562085] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145707, 'name': ReconfigVM_Task, 'duration_secs': 0.912169} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.564150] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.564735] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Reconfigured VM instance instance-00000012 to attach disk [datastore2] 22899269-28fc-446e-9533-1f6862a9e8db/22899269-28fc-446e-9533-1f6862a9e8db.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 684.565667] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 684.565667] env[66641]: value = "task-5145710" [ 684.565667] env[66641]: _type = "Task" [ 684.565667] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.569212] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.569212] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0f2eec8-3faf-4da1-9563-0beae3ee855d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.572037] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.573019] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 684.576047] env[66641]: DEBUG oslo_vmware.api [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Task: {'id': task-5145708, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.524212} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.576470] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 684.578783] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 684.579241] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 684.580024] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.580024] env[66641]: INFO nova.compute.manager [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Took 1.70 seconds to destroy the instance on the hypervisor. [ 684.580209] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 684.581380] env[66641]: DEBUG nova.compute.manager [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 684.581674] env[66641]: DEBUG nova.network.neutron [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 684.582329] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.582777] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.596604] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.597548] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 684.597548] env[66641]: value = "task-5145711" [ 684.597548] env[66641]: _type = "Task" [ 684.597548] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.610301] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145711, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.868739] env[66641]: DEBUG nova.compute.utils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 684.878365] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 684.878934] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 684.884086] env[66641]: WARNING neutronclient.v2_0.client [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.884086] env[66641]: WARNING neutronclient.v2_0.client [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 684.884086] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 684.884086] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 684.895190] env[66641]: DEBUG oslo_concurrency.lockutils [None req-00543dc6-fefa-49c9-80fa-73dc277f282d tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 0.531s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 684.895898] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.467s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 684.900597] env[66641]: INFO nova.compute.claims [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.079401] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.082471] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 685.108258] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145711, 'name': Rename_Task, 'duration_secs': 0.265424} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.108559] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 685.108811] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c469fb78-28fb-4493-af60-51d9d9ae4ed2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.116281] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 685.116281] env[66641]: value = "task-5145712" [ 685.116281] env[66641]: _type = "Task" [ 685.116281] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.126774] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.333838] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 685.333838] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b4dd390-71af-4d63-8ccb-837b46f50467 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.341255] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 685.341255] env[66641]: value = "task-5145713" [ 685.341255] env[66641]: _type = "Task" [ 685.341255] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.351053] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145713, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.379454] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 685.579922] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145710, 'name': ReconfigVM_Task, 'duration_secs': 0.640997} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.580310] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Reconfigured VM instance instance-00000011 to attach disk [datastore2] fb1bf80c-b243-4c25-9aaf-c459ca926090/fb1bf80c-b243-4c25-9aaf-c459ca926090.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.580951] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dac421ca-9fff-4c53-bd3a-a45af979b2e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.588037] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 685.588037] env[66641]: value = "task-5145714" [ 685.588037] env[66641]: _type = "Task" [ 685.588037] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.596986] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145714, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.627506] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.794492] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Received event network-changed-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 685.794800] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Refreshing instance network info cache due to event network-changed-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 685.795113] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquiring lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.795264] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquired lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 685.795424] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Refreshing network info cache for port f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 685.854025] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145713, 'name': PowerOffVM_Task, 'duration_secs': 0.195036} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.855681] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 685.856751] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 685.856751] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426cd1f1-975a-4bbf-ae1b-40548f686a72 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.865173] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 685.866051] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc5ab63e-ca6a-464e-99a3-138b1c3d42e6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.895389] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 685.895709] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 685.896068] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Deleting the datastore file [datastore2] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 685.896321] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1ceeed67-86b0-4780-8252-9f682a903d96 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.905129] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 685.905129] env[66641]: value = "task-5145716" [ 685.905129] env[66641]: _type = "Task" [ 685.905129] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.916699] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.977201] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.977555] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 685.990707] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 685.990707] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.001735] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Successfully updated port: 653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 686.035560] env[66641]: DEBUG nova.policy [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbd14493616a438491862fb37f9d3fd7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ef1e42e42cd492d8dce962e90011df8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 686.105681] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145714, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.127949] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.155510] env[66641]: INFO nova.compute.manager [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Rebuilding instance [ 686.210428] env[66641]: DEBUG nova.compute.manager [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 686.212143] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3eb17fa-8fd5-4157-8033-f3fa10e0d7ec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.216596] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482266ea-a105-4700-8e4f-6e6e88c212df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.223513] env[66641]: DEBUG oslo_concurrency.lockutils [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 686.223862] env[66641]: DEBUG oslo_concurrency.lockutils [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 686.223907] env[66641]: DEBUG nova.compute.manager [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 686.229464] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d948c96f-f731-4129-ab22-06156e5ca291 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.232757] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373cce82-4e41-4512-8c39-13ee81eb8aab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.241054] env[66641]: DEBUG nova.compute.manager [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=66641) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 686.241486] env[66641]: DEBUG nova.objects.instance [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lazy-loading 'flavor' on Instance uuid 695d1824-cf34-46c1-b623-b474d2f1f8ef {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 686.273557] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d8be5f2-88bb-4731-8c33-0119d812bfa3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.281995] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ba7dcf8-6cce-46ef-b434-9f3507cb5e95 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.297753] env[66641]: DEBUG nova.compute.provider_tree [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.299768] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 686.300012] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 686.390049] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 686.417630] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.339606} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 686.419828] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 686.420056] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 686.420228] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 686.420412] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 686.420551] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 686.420690] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 686.420890] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 686.421067] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 686.421217] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 686.421373] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 686.421720] env[66641]: DEBUG nova.virt.hardware [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 686.421806] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 686.421968] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 686.422200] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.425095] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ffba43-d168-475b-95ec-9a670e722286 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.433305] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744cd8f8-1998-4b81-b949-8384a4955745 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.505618] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.505894] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 686.506187] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 686.602349] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145714, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.628366] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.798538] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Successfully created port: 33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 686.808715] env[66641]: DEBUG nova.scheduler.client.report [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 687.012113] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.014679] env[66641]: WARNING openstack [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.028518] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.029823] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.039174] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 687.104231] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145714, 'name': Rename_Task, 'duration_secs': 1.337923} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.104513] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 687.104766] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9f300697-74f4-4a2d-beb7-b49550d86029 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.111240] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 687.111240] env[66641]: value = "task-5145717" [ 687.111240] env[66641]: _type = "Task" [ 687.111240] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.119470] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.128722] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.200159] env[66641]: DEBUG nova.network.neutron [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 687.237863] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 687.240526] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f919b5e9-39d0-486b-b354-dbcf5c787d1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.249327] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 687.249327] env[66641]: value = "task-5145718" [ 687.249327] env[66641]: _type = "Task" [ 687.249327] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.259319] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145718, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.279350] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 687.279553] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c897b772-f6ce-4e68-87fb-233e7a2ed3a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.290861] env[66641]: DEBUG oslo_vmware.api [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 687.290861] env[66641]: value = "task-5145719" [ 687.290861] env[66641]: _type = "Task" [ 687.290861] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.303548] env[66641]: DEBUG oslo_vmware.api [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145719, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.314173] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 687.314859] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 687.318955] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.564s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 687.320396] env[66641]: INFO nova.compute.claims [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.473833] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 687.474486] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 687.474895] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 687.475286] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 687.475657] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 687.475940] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 687.476872] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.476872] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 687.476872] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 687.477512] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 687.478608] env[66641]: DEBUG nova.virt.hardware [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 687.482357] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-909fbd8b-989a-4d8c-853e-b292c6ea37a1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.496337] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1c60280-d52d-4243-98cc-802c09b97063 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.514396] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 687.522628] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 687.523080] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 687.523214] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-633068e5-37a3-411b-b0d5-352ae1f39790 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.544834] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 687.544834] env[66641]: value = "task-5145720" [ 687.544834] env[66641]: _type = "Task" [ 687.544834] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.553863] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145720, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.627851] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145717, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.633802] env[66641]: DEBUG oslo_vmware.api [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145712, 'name': PowerOnVM_Task, 'duration_secs': 2.305572} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.634088] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 687.634300] env[66641]: INFO nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Took 10.04 seconds to spawn the instance on the hypervisor. [ 687.634487] env[66641]: DEBUG nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 687.635291] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c18aac-a56c-4452-b2a4-62a41f8b21d4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.706754] env[66641]: INFO nova.compute.manager [-] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Took 3.12 seconds to deallocate network for instance. [ 687.760203] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145718, 'name': PowerOffVM_Task, 'duration_secs': 0.275383} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.760491] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 687.760709] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 687.761568] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a1b730a-5b1c-4470-a70c-e4c81a55fae4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.769105] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 687.769361] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc7156c7-00af-4b89-afd1-25f6a3af9274 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.803300] env[66641]: DEBUG oslo_vmware.api [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145719, 'name': PowerOffVM_Task, 'duration_secs': 0.299757} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.803514] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 687.803735] env[66641]: DEBUG nova.compute.manager [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 687.804688] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6dfb9-59c4-4147-9fe4-a17efca2043f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.825199] env[66641]: DEBUG nova.compute.utils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 687.829917] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 687.830490] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 687.830971] env[66641]: WARNING neutronclient.v2_0.client [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.831493] env[66641]: WARNING neutronclient.v2_0.client [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 687.832471] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 687.833177] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 687.845016] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 687.848654] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 687.848879] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 687.849094] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleting the datastore file [datastore2] fe143d94-57ce-46d4-85e3-80d84af66dfc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.849945] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11f71a56-5192-42ef-aaf6-0a9f71778d22 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.862552] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 687.862552] env[66641]: value = "task-5145722" [ 687.862552] env[66641]: _type = "Task" [ 687.862552] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.874312] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.058890] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145720, 'name': CreateVM_Task, 'duration_secs': 0.438156} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.059311] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 688.060024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.060425] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.060917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 688.061437] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31852538-ddd8-4449-92df-2407bdc56baa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.070375] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 688.070375] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e27779-cf4d-7f49-9d73-800ec09d07ff" [ 688.070375] env[66641]: _type = "Task" [ 688.070375] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.086388] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e27779-cf4d-7f49-9d73-800ec09d07ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.124826] env[66641]: DEBUG oslo_vmware.api [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145717, 'name': PowerOnVM_Task, 'duration_secs': 0.904244} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.124826] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 688.124826] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Took 12.89 seconds to spawn the instance on the hypervisor. [ 688.124962] env[66641]: DEBUG nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 688.125849] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-327cdd0f-31fc-4398-88d2-c58314566deb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.154318] env[66641]: INFO nova.compute.manager [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Took 31.91 seconds to build instance. [ 688.213650] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 688.264827] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.265399] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.318814] env[66641]: DEBUG oslo_concurrency.lockutils [None req-071adade-7508-448d-a88d-7a7dac9b4e9c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.095s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.375402] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.421829} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.375627] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.375807] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 688.375974] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.410869] env[66641]: DEBUG nova.policy [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8a66322faba64e1088266a2f36ae69f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '42e99a1d4fdb478d825190d12850ad89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 688.588682] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e27779-cf4d-7f49-9d73-800ec09d07ff, 'name': SearchDatastore_Task, 'duration_secs': 0.026587} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.589704] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 688.589704] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 688.589704] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.589954] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 688.590046] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 688.590709] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7f12c16-4c21-4c05-a113-18c95fcd0bb3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.602390] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 688.602603] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 688.603399] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5130787a-9136-4522-b9ee-84bbad0e379e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.609733] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 688.609733] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52afc8db-2ca1-354a-8e77-93132863944c" [ 688.609733] env[66641]: _type = "Task" [ 688.609733] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 688.622841] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52afc8db-2ca1-354a-8e77-93132863944c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 688.642939] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0249d5aa-bf04-499e-bb37-13c807f80339 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.651184] env[66641]: INFO nova.compute.manager [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Took 32.60 seconds to build instance. [ 688.655389] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b89adf36-4ed2-48d0-9f46-7ca182bed3d4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.659697] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e26c4ad8-bb45-4d75-8227-8e71e9ce5552 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.437s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 688.690895] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-800574f1-cac5-4c8e-84a9-bf6db229e72a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.700709] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d7013b7-e428-40ce-aee7-1bd67be5f9c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.717285] env[66641]: DEBUG nova.compute.provider_tree [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.734822] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Successfully updated port: 33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 688.858633] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 688.904823] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 688.905642] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 688.906437] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 688.906677] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 688.906868] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 688.907034] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 688.907271] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.907615] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 688.907936] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 688.908283] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 688.908580] env[66641]: DEBUG nova.virt.hardware [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 688.909548] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4c4ea51-5c0e-4b89-80ff-99720b4bc568 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.919686] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89e37355-7e59-42dc-8446-91ebab3ca57b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.925848] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 688.926287] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 688.990561] env[66641]: DEBUG nova.network.neutron [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Updated VIF entry in instance network info cache for port 2fd71b07-2d9b-426a-a09a-715cea2538e3. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 688.990940] env[66641]: DEBUG nova.network.neutron [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Updating instance_info_cache with network_info: [{"id": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "address": "fa:16:3e:3f:01:c2", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2fd71b07-2d", "ovs_interfaceid": "2fd71b07-2d9b-426a-a09a-715cea2538e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 689.126712] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52afc8db-2ca1-354a-8e77-93132863944c, 'name': SearchDatastore_Task, 'duration_secs': 0.037383} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.127599] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d6e1ba0-560f-4e30-82fa-439e0d9ba952 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.134671] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 689.134671] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52457863-6f93-1234-71cb-a2f687e92c6e" [ 689.134671] env[66641]: _type = "Task" [ 689.134671] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.144967] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52457863-6f93-1234-71cb-a2f687e92c6e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.153593] env[66641]: DEBUG oslo_concurrency.lockutils [None req-45ce4dd4-d896-46bf-ac48-4c50debd8115 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.143s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.220467] env[66641]: DEBUG nova.scheduler.client.report [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 689.237719] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.237912] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquired lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 689.241088] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 689.392209] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.392662] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.411037] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Successfully created port: d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 689.423855] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 689.424269] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 689.424363] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 689.424690] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 689.424910] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 689.425103] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 689.425622] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.425834] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 689.426098] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 689.426308] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 689.426518] env[66641]: DEBUG nova.virt.hardware [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 689.427477] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944cfb33-b857-4c71-82b8-6d8a2adc4196 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.436874] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a64da9-33c3-4f7c-828a-453d43496d40 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.453617] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d8:c3:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 689.460949] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 689.461311] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 689.461534] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8efa35d1-82db-4a0e-8c20-806f6acb229e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.483412] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 689.483412] env[66641]: value = "task-5145723" [ 689.483412] env[66641]: _type = "Task" [ 689.483412] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.491591] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145723, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.493935] env[66641]: DEBUG oslo_concurrency.lockutils [req-03f2b201-fec4-41c6-9930-6bc5d1dc8784 req-14db5cf8-0a04-4b17-bf20-979b2e4cb660 service nova] Releasing lock "refresh_cache-434e66b1-5ea7-43d2-9b41-bdeda2a312d7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.564064] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.564064] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.647262] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52457863-6f93-1234-71cb-a2f687e92c6e, 'name': SearchDatastore_Task, 'duration_secs': 0.030142} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 689.647693] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 689.648124] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 689.648568] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc8da069-f5d5-4bfc-9a24-f15791f9ad29 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.660021] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 689.660021] env[66641]: value = "task-5145724" [ 689.660021] env[66641]: _type = "Task" [ 689.660021] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 689.669755] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 689.727075] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.409s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.729728] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 689.735447] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.653s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.735447] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 689.735785] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 689.738517] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.523s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 689.738927] env[66641]: DEBUG nova.objects.instance [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lazy-loading 'resources' on Instance uuid e7bfb26e-b326-4a39-8f20-79dbbfb74eb8 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 689.741665] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8828e371-ed52-4f32-a997-30bed2216637 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.747754] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 689.748489] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 689.763108] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 689.782956] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f861900-7afd-4e07-bf04-41ea43c7f916 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.802803] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3df01e-09e2-4548-9dbe-173d65d88fb7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.814165] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397d34c5-c271-4fa5-8ac7-0389e25dd4ae {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.851766] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180054MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 689.851928] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.000215] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145723, 'name': CreateVM_Task, 'duration_secs': 0.43761} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.003707] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 690.005371] env[66641]: WARNING openstack [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.005870] env[66641]: WARNING openstack [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.012503] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.012503] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.012824] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 690.013717] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-272a2c22-b937-40c8-8122-280d7aa56d09 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.023285] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 690.023285] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5253f07b-6300-93bd-9002-79d351647a99" [ 690.023285] env[66641]: _type = "Task" [ 690.023285] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.036087] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5253f07b-6300-93bd-9002-79d351647a99, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.091705] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00adb7c5-e0ac-4893-a3f3-d4b11e131d49 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.105904] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31338b99-6fe1-43d6-aad8-d8ecee7819df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.155812] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a13b3d57-96c8-4792-893c-dfdb198907e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.167695] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Updated VIF entry in instance network info cache for port f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 690.168260] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Updating instance_info_cache with network_info: [{"id": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "address": "fa:16:3e:26:f9:99", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3fb5c31-b4", "ovs_interfaceid": "f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.175622] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc29998a-ed8f-4b33-b4d6-34ddebf3d3ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.184327] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145724, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.198173] env[66641]: DEBUG nova.compute.provider_tree [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.242563] env[66641]: DEBUG nova.compute.utils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 690.244470] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 690.244859] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 690.245310] env[66641]: WARNING neutronclient.v2_0.client [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.245761] env[66641]: WARNING neutronclient.v2_0.client [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 690.246546] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.246938] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.524936] env[66641]: DEBUG nova.network.neutron [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Updating instance_info_cache with network_info: [{"id": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "address": "fa:16:3e:81:45:1f", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap653c0e4c-f1", "ovs_interfaceid": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 690.542691] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5253f07b-6300-93bd-9002-79d351647a99, 'name': SearchDatastore_Task, 'duration_secs': 0.063249} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.544346] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 690.544808] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 690.553300] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.555867] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.555867] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.555867] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.555867] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.556279] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-137fabcd-5621-4893-939a-4b95db5e2fbd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.571105] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.571105] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 690.571105] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d34cd808-1579-4a23-b53b-fd6145ae48fa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.577842] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 690.577842] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520a9897-a2bb-7a7c-20e3-3a2d13c1bc9c" [ 690.577842] env[66641]: _type = "Task" [ 690.577842] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.588590] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520a9897-a2bb-7a7c-20e3-3a2d13c1bc9c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.672328] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692236} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.672639] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 690.672913] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 690.673185] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a85986ce-8df9-4779-8978-36bd317253ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.675991] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Releasing lock "refresh_cache-d14528cb-f26e-4fe5-8bf9-34900571fe03" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 690.676241] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Received event network-vif-plugged-dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 690.676439] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquiring lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 690.676651] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 690.676788] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 690.676939] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] No waiting events found dispatching network-vif-plugged-dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 690.677113] env[66641]: WARNING nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Received unexpected event network-vif-plugged-dbb7738c-a759-4460-8f36-9f673e12b204 for instance with vm_state building and task_state spawning. [ 690.677319] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Received event network-changed-dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 690.677422] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Refreshing instance network info cache due to event network-changed-dbb7738c-a759-4460-8f36-9f673e12b204. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 690.677595] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquiring lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.677727] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquired lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 690.677880] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Refreshing network info cache for port dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 690.685205] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 690.685205] env[66641]: value = "task-5145725" [ 690.685205] env[66641]: _type = "Task" [ 690.685205] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.702641] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145725, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.706123] env[66641]: DEBUG nova.scheduler.client.report [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 690.757401] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 691.030993] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.031393] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Instance network_info: |[{"id": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "address": "fa:16:3e:81:45:1f", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap653c0e4c-f1", "ovs_interfaceid": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 691.031852] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:45:1f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '653c0e4c-f1b9-44f8-abcc-6d580fd70565', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 691.042644] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Creating folder: Project (2ab82c335b15478788a818319bf11c03). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.042781] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e0c8b9ae-c3b8-4a6d-8a56-168602718e39 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.056982] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Created folder: Project (2ab82c335b15478788a818319bf11c03) in parent group-v1000566. [ 691.056982] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Creating folder: Instances. Parent ref: group-v1000619. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 691.056982] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c3ce5507-6c4d-4d75-b9fb-da3057b5f45b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.068796] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Created folder: Instances in parent group-v1000619. [ 691.069073] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 691.069294] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 691.069514] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-37a81559-cc0c-4344-87b7-0f64067fe6b4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.101142] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520a9897-a2bb-7a7c-20e3-3a2d13c1bc9c, 'name': SearchDatastore_Task, 'duration_secs': 0.012478} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.103411] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 691.103411] env[66641]: value = "task-5145728" [ 691.103411] env[66641]: _type = "Task" [ 691.103411] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.103593] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2f72c8b-9b7b-4459-9d0f-31f298a2ab39 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.116138] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145728, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.116766] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 691.116766] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9b058-2a57-85cc-28d5-ef6e043d5e7f" [ 691.116766] env[66641]: _type = "Task" [ 691.116766] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.127134] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9b058-2a57-85cc-28d5-ef6e043d5e7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.181410] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.181923] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.200770] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145725, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072673} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.201051] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 691.202066] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb915cca-6d58-472e-8878-c10295195a5b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.216480] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.480s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.230039] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfiguring VM instance instance-0000000b to attach disk [datastore1] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 691.234558] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Successfully updated port: d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 691.236200] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.384s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 691.238977] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-64d8b083-2742-45f0-9a55-9b1bd5395cd6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.267940] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 691.267940] env[66641]: value = "task-5145729" [ 691.267940] env[66641]: _type = "Task" [ 691.267940] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.283904] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145729, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.285094] env[66641]: INFO nova.scheduler.client.report [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Deleted allocations for instance e7bfb26e-b326-4a39-8f20-79dbbfb74eb8 [ 691.296041] env[66641]: DEBUG nova.policy [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03e9406a462042148d6d1a926d0bbab9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '95981b554a7f41e189993a89ea395b43', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 691.604131] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.604131] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.630997] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145728, 'name': CreateVM_Task, 'duration_secs': 0.499976} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.634755] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 691.635173] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9b058-2a57-85cc-28d5-ef6e043d5e7f, 'name': SearchDatastore_Task, 'duration_secs': 0.014446} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.636589] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 691.636708] env[66641]: WARNING openstack [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 691.643608] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.643608] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.643608] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 691.643870] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 691.644361] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 691.644744] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cb2c178-f772-4068-aa51-a2b0638691f8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.647696] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1afd637-67ea-4123-95b2-12f7bdfb0a03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.660489] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 691.660489] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5294a638-b6a8-5015-0561-7e6f2c4bb7f7" [ 691.660489] env[66641]: _type = "Task" [ 691.660489] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.663199] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 691.663199] env[66641]: value = "task-5145730" [ 691.663199] env[66641]: _type = "Task" [ 691.663199] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.679888] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5294a638-b6a8-5015-0561-7e6f2c4bb7f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.686784] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.761477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.761854] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquired lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 691.761854] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 691.775459] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 691.784214] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145729, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.793349] env[66641]: DEBUG oslo_concurrency.lockutils [None req-5bbe34c0-e5a9-4ed7-97cd-6c9d04ce59cb tempest-ServerExternalEventsTest-185443793 tempest-ServerExternalEventsTest-185443793-project-member] Lock "e7bfb26e-b326-4a39-8f20-79dbbfb74eb8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.417s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 691.811474] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 691.811730] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 691.811874] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 691.812054] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 691.812190] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 691.812700] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 691.812700] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.812794] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 691.812932] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 691.813095] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 691.813259] env[66641]: DEBUG nova.virt.hardware [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 691.814679] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2e8433-93eb-45be-84b8-3f22357d3bec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.827232] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a84f544f-a11f-43f8-96ca-368255013275 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.998481] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Successfully created port: 932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 692.178882] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.185093] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5294a638-b6a8-5015-0561-7e6f2c4bb7f7, 'name': SearchDatastore_Task, 'duration_secs': 0.022623} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.185530] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 692.185869] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 692.186239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.186469] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 692.186782] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 692.187163] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3cc254af-081b-486f-a85f-1dcf6a559a3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.204042] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 692.204042] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 692.204671] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7befee9f-d28e-4233-9d96-8b840f4d8f7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.212763] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 692.212763] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285249d-57e1-d1b4-ad53-b07c885eb1c6" [ 692.212763] env[66641]: _type = "Task" [ 692.212763] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.221230] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285249d-57e1-d1b4-ad53-b07c885eb1c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.248903] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Applying migration context for instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 as it has an incoming, in-progress migration 1ed74871-41c7-4c89-bbf4-b55c237571e3. Migration status is finished {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 692.250838] env[66641]: INFO nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating resource usage from migration 1ed74871-41c7-4c89-bbf4-b55c237571e3 [ 692.268760] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.269223] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.275620] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 692.282447] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.282640] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance fe143d94-57ce-46d4-85e3-80d84af66dfc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.282715] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283053] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 695d1824-cf34-46c1-b623-b474d2f1f8ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283053] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Migration 1ed74871-41c7-4c89-bbf4-b55c237571e3 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 692.283208] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283208] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 434e66b1-5ea7-43d2-9b41-bdeda2a312d7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283351] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance d14528cb-f26e-4fe5-8bf9-34900571fe03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283398] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance fb1bf80c-b243-4c25-9aaf-c459ca926090 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283667] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 22899269-28fc-446e-9533-1f6862a9e8db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283667] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance bc13bbe2-6e02-4d98-9e50-94f772d89ac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283918] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance ac5bb5a2-e630-42f0-be0c-24f52be81367 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.283918] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance a2d567c9-09af-44e1-aad2-c8f394a09714 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.284082] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance eeb2556a-4e0c-43d7-83bd-942be5d5fdd2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 692.284277] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 14 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 692.284494] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3264MB phys_disk=100GB used_disk=14GB total_vcpus=48 used_vcpus=14 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '13', 'num_vm_active': '7', 'num_task_None': '7', 'num_os_type_None': '13', 'num_proj_417e1e40aca447229001fa725c82e9ca': '1', 'io_workload': '4', 'num_vm_resized': '1', 'num_proj_6624f269bb8840e98c10259f006ce962': '1', 'num_task_rebuild_spawning': '2', 'num_proj_8449c95dd0f74492929e4d1d492ac9d2': '1', 'num_proj_7f8e91d542844e1fad6ba8c274d0756d': '1', 'num_vm_stopped': '1', 'num_proj_17792883c6504d0f9f91fb6beb25b087': '1', 'num_proj_196b2ed18f7142a18f9093869dadb46e': '3', 'num_proj_82383b2f8cf2490ebdf92495d9dc20e6': '1', 'num_vm_building': '4', 'num_task_spawning': '3', 'num_proj_2ab82c335b15478788a818319bf11c03': '1', 'num_proj_8ef1e42e42cd492d8dce962e90011df8': '1', 'num_proj_42e99a1d4fdb478d825190d12850ad89': '1', 'num_task_block_device_mapping': '1', 'num_proj_95981b554a7f41e189993a89ea395b43': '1'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 692.295803] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145729, 'name': ReconfigVM_Task, 'duration_secs': 0.583813} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.296110] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Reconfigured VM instance instance-0000000b to attach disk [datastore1] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763/5963a8ba-9d2c-42c3-a5da-25c29bf9e763.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 692.296969] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85c50a6a-fa9f-435d-b54c-ed3ec861e540 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.309874] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 692.309874] env[66641]: value = "task-5145731" [ 692.309874] env[66641]: _type = "Task" [ 692.309874] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.326407] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145731, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.553755] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8da80c98-7458-43a1-9489-9bf57ad39e81 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.563194] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b09dd470-983e-407a-a891-cd5c62f9ab9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.571023] env[66641]: DEBUG nova.network.neutron [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Updating instance_info_cache with network_info: [{"id": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "address": "fa:16:3e:fc:ab:bb", "network": {"id": "114a8269-be9e-4964-abfe-24aee4c612cd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1298879401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8ef1e42e42cd492d8dce962e90011df8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33cb5bcf-2e", "ovs_interfaceid": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 692.608078] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 692.608078] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 692.615234] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd031f97-98d1-4f1c-ba99-3d6b4bad075c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.625035] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dea1295f-84ce-42d3-bfc7-ae71df1415fe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.640432] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.678353] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145730, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.962878} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.678631] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 692.678836] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.679143] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-70368992-6ac1-4348-a777-45e1a36fbc3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.687105] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 692.687105] env[66641]: value = "task-5145732" [ 692.687105] env[66641]: _type = "Task" [ 692.687105] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.696810] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145732, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.728318] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285249d-57e1-d1b4-ad53-b07c885eb1c6, 'name': SearchDatastore_Task, 'duration_secs': 0.01025} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.729476] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4622330-151c-42ce-a343-b9a828ba2f4f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.735837] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 692.735837] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521eb28c-15da-67c4-3c53-d19729613115" [ 692.735837] env[66641]: _type = "Task" [ 692.735837] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.744285] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521eb28c-15da-67c4-3c53-d19729613115, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.810132] env[66641]: DEBUG nova.compute.manager [req-ca17bdd8-6d32-4227-9739-15c5c580d960 req-8b1b35f1-d811-4408-9e0a-02bd5405a644 service nova] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Received event network-vif-deleted-d3f790d9-7633-46ba-842e-725e9680d0ea {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 692.826157] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145731, 'name': Rename_Task, 'duration_secs': 0.389971} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.826157] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 692.826157] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5707d43-7a3d-42c3-a898-b96bb45b4c77 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.833138] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Waiting for the task: (returnval){ [ 692.833138] env[66641]: value = "task-5145733" [ 692.833138] env[66641]: _type = "Task" [ 692.833138] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.841596] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145733, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.075519] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Releasing lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.076088] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Instance network_info: |[{"id": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "address": "fa:16:3e:fc:ab:bb", "network": {"id": "114a8269-be9e-4964-abfe-24aee4c612cd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1298879401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "8ef1e42e42cd492d8dce962e90011df8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33cb5bcf-2e", "ovs_interfaceid": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 693.076626] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:ab:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7654928b-7afe-42e3-a18d-68ecc775cefe', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 693.087618] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Creating folder: Project (8ef1e42e42cd492d8dce962e90011df8). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.088647] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac614c36-16b3-4640-a14f-283344316b2b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.106591] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Created folder: Project (8ef1e42e42cd492d8dce962e90011df8) in parent group-v1000566. [ 693.106591] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Creating folder: Instances. Parent ref: group-v1000622. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 693.106591] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52f1d85f-d797-48f1-9360-d543a4cea46b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.119238] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Created folder: Instances in parent group-v1000622. [ 693.119482] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 693.119695] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 693.119907] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-28e6fdf2-2bfd-48d0-8695-b7bcdfdafab7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.864713] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.865198] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.873850] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 693.874222] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 693.883928] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Successfully updated port: 932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 693.886475] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 693.893790] env[66641]: DEBUG nova.compute.manager [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Received event network-vif-plugged-653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 693.893982] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Acquiring lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 693.894197] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 693.894348] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 693.894503] env[66641]: DEBUG nova.compute.manager [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] No waiting events found dispatching network-vif-plugged-653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 693.894779] env[66641]: WARNING nova.compute.manager [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Received unexpected event network-vif-plugged-653c0e4c-f1b9-44f8-abcc-6d580fd70565 for instance with vm_state building and task_state spawning. [ 693.894779] env[66641]: DEBUG nova.compute.manager [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Received event network-changed-653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 693.894929] env[66641]: DEBUG nova.compute.manager [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Refreshing instance network info cache due to event network-changed-653c0e4c-f1b9-44f8-abcc-6d580fd70565. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 693.895111] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Acquiring lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.895545] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Acquired lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 693.895545] env[66641]: DEBUG nova.network.neutron [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Refreshing network info cache for port 653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 693.918603] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145732, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075986} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.925782] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 693.926115] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 693.926115] env[66641]: value = "task-5145736" [ 693.926115] env[66641]: _type = "Task" [ 693.926115] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.926897] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]521eb28c-15da-67c4-3c53-d19729613115, 'name': SearchDatastore_Task, 'duration_secs': 0.012034} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.928721] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c8d3e9-d6b8-4e76-b511-14a5254bce49 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.932317] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 693.932519] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] bc13bbe2-6e02-4d98-9e50-94f772d89ac0/bc13bbe2-6e02-4d98-9e50-94f772d89ac0.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 693.940073] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-836b198c-beed-4d0d-aa26-e4e2c39ba0bd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.942644] env[66641]: DEBUG oslo_vmware.api [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Task: {'id': task-5145733, 'name': PowerOnVM_Task, 'duration_secs': 0.60002} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.943932] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 693.944149] env[66641]: DEBUG nova.compute.manager [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 693.945539] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54f0fdff-a0f3-404d-a402-17ce7d5056a1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.970955] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145736, 'name': CreateVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.971582] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 693.971582] env[66641]: value = "task-5145737" [ 693.971582] env[66641]: _type = "Task" [ 693.971582] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.988326] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 693.990799] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8f92a371-7ef0-4b6c-8818-8e952894335e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.039022] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145737, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.039022] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 694.039022] env[66641]: value = "task-5145738" [ 694.039022] env[66641]: _type = "Task" [ 694.039022] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.048399] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145738, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.324327] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Updated VIF entry in instance network info cache for port dbb7738c-a759-4460-8f36-9f673e12b204. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 694.324753] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Updating instance_info_cache with network_info: [{"id": "dbb7738c-a759-4460-8f36-9f673e12b204", "address": "fa:16:3e:8d:c8:a4", "network": {"id": "1a1b0407-5bdd-44d8-98be-0152d66c949c", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-1282466236-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196b2ed18f7142a18f9093869dadb46e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f4f4d0bb-61eb-4597-bc00-c9fdbc85f93d", "external-id": "nsx-vlan-transportzone-470", "segmentation_id": 470, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbb7738c-a7", "ovs_interfaceid": "dbb7738c-a759-4460-8f36-9f673e12b204", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 694.340997] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.341592] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.401976] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.402236] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquired lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 694.402413] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 694.407027] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 694.407027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.168s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.446440] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.448490] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.471029] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145736, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.504251] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145737, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.546739] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.547216] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.547532] env[66641]: DEBUG nova.objects.instance [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 694.563954] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145738, 'name': ReconfigVM_Task, 'duration_secs': 0.424559} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.564554] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Reconfigured VM instance instance-00000009 to attach disk [datastore1] fe143d94-57ce-46d4-85e3-80d84af66dfc/fe143d94-57ce-46d4-85e3-80d84af66dfc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 694.567128] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c78e16d9-d7a2-4b21-8499-ee74f66375b5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.577322] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 694.577322] env[66641]: value = "task-5145739" [ 694.577322] env[66641]: _type = "Task" [ 694.577322] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.592029] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145739, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.662979] env[66641]: DEBUG nova.network.neutron [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Updating instance_info_cache with network_info: [{"id": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "address": "fa:16:3e:13:e7:e4", "network": {"id": "18515ee6-2926-44bf-b2bc-d5101fd6de51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1940026041-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "42e99a1d4fdb478d825190d12850ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd66320d7-48", "ovs_interfaceid": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 694.827569] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Releasing lock "refresh_cache-fb1bf80c-b243-4c25-9aaf-c459ca926090" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 694.828813] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Received event network-vif-plugged-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 694.828813] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquiring lock "22899269-28fc-446e-9533-1f6862a9e8db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 694.829186] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Lock "22899269-28fc-446e-9533-1f6862a9e8db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 694.829529] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Lock "22899269-28fc-446e-9533-1f6862a9e8db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 694.829973] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] No waiting events found dispatching network-vif-plugged-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 694.831898] env[66641]: WARNING nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Received unexpected event network-vif-plugged-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a for instance with vm_state building and task_state spawning. [ 694.832202] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Received event network-changed-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 694.833612] env[66641]: DEBUG nova.compute.manager [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Refreshing instance network info cache due to event network-changed-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 694.833612] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquiring lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.833612] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Acquired lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 694.833612] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Refreshing network info cache for port f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 694.894850] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.895269] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.906303] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.906710] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.912494] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 694.949957] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145736, 'name': CreateVM_Task, 'duration_secs': 0.599245} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.950230] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 694.951888] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.951888] env[66641]: WARNING openstack [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.957736] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.957900] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 694.958248] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 694.963492] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7830b587-f2e4-47f5-b9fc-9330a45f26c0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.966505] env[66641]: DEBUG nova.compute.manager [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 694.967531] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6266f2a-3306-4af4-9fef-9f25f3adabe9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.984844] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 694.985241] env[66641]: WARNING openstack [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 694.995732] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 694.995732] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ca3cd1-049e-aaaa-efa8-a06168be9662" [ 694.995732] env[66641]: _type = "Task" [ 694.995732] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.002545] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.002928] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.025133] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145737, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649312} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.025637] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] bc13bbe2-6e02-4d98-9e50-94f772d89ac0/bc13bbe2-6e02-4d98-9e50-94f772d89ac0.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 695.025851] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 695.026121] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf05a138-b00f-47ec-8207-25d41b8e2b38 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.035188] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ca3cd1-049e-aaaa-efa8-a06168be9662, 'name': SearchDatastore_Task, 'duration_secs': 0.011603} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.035188] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.035308] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 695.035532] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.035668] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 695.035840] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 695.036982] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a188eaad-8a66-4c6f-baa6-f90bc3139e1e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.043938] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 695.043938] env[66641]: value = "task-5145740" [ 695.043938] env[66641]: _type = "Task" [ 695.043938] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.059670] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145740, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.061520] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 695.061810] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 695.068625] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87af527e-ace6-49bc-bab7-b9670543b12a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.079317] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 695.079317] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f009e8-4b70-2906-e14a-1bc1b044f10e" [ 695.079317] env[66641]: _type = "Task" [ 695.079317] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.097020] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145739, 'name': Rename_Task, 'duration_secs': 0.242415} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.097020] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 695.097365] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f009e8-4b70-2906-e14a-1bc1b044f10e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.101643] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5cbc8cd5-8bf6-4135-a8ba-1e640f03988a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.111133] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 695.111133] env[66641]: value = "task-5145741" [ 695.111133] env[66641]: _type = "Task" [ 695.111133] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.121499] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.145097] env[66641]: DEBUG nova.network.neutron [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Updated VIF entry in instance network info cache for port 653c0e4c-f1b9-44f8-abcc-6d580fd70565. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 695.145097] env[66641]: DEBUG nova.network.neutron [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Updating instance_info_cache with network_info: [{"id": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "address": "fa:16:3e:81:45:1f", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.87", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap653c0e4c-f1", "ovs_interfaceid": "653c0e4c-f1b9-44f8-abcc-6d580fd70565", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.163518] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.163915] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.175893] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Releasing lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.178056] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Instance network_info: |[{"id": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "address": "fa:16:3e:13:e7:e4", "network": {"id": "18515ee6-2926-44bf-b2bc-d5101fd6de51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1940026041-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "42e99a1d4fdb478d825190d12850ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd66320d7-48", "ovs_interfaceid": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 695.178289] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:e7:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b356db78-99c7-4464-822c-fc7e193f7878', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd66320d7-4884-46e8-83ab-a8a6fd0b09a7', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.187287] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Creating folder: Project (42e99a1d4fdb478d825190d12850ad89). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.187287] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb28c752-8228-4c59-97bb-20332b21577a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.199876] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Created folder: Project (42e99a1d4fdb478d825190d12850ad89) in parent group-v1000566. [ 695.201321] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Creating folder: Instances. Parent ref: group-v1000625. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.201321] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-26adcd22-bb40-4509-838e-137748dc7a29 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.214572] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Created folder: Instances in parent group-v1000625. [ 695.214862] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 695.218020] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 695.218020] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e05aea79-8509-41de-81df-2f6fd21ec755 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.239682] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.239682] env[66641]: value = "task-5145744" [ 695.239682] env[66641]: _type = "Task" [ 695.239682] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.257140] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145744, 'name': CreateVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.308696] env[66641]: DEBUG nova.network.neutron [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Updating instance_info_cache with network_info: [{"id": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "address": "fa:16:3e:6f:e9:79", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.49", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932f4137-a3", "ovs_interfaceid": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 695.340022] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.340022] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.454313] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.454313] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.500989] env[66641]: INFO nova.compute.manager [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] instance snapshotting [ 695.513542] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddf58bc-d710-4899-8268-79c2ea08a4ad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.533581] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 695.533693] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 695.533779] env[66641]: DEBUG nova.compute.manager [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Going to confirm migration 1 {{(pid=66641) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 695.539746] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9aa7b15-94cb-4b3e-ab97-77d91cd1ffa4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.558104] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145740, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0768} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.558390] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.559225] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2bbb51-df84-4faa-b88a-f5f4a91c0e82 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.562746] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1dc7a723-325a-4573-a274-c1b14bb6ad87 tempest-ServersAdmin275Test-1758624687 tempest-ServersAdmin275Test-1758624687-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 695.588071] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Reconfiguring VM instance instance-00000013 to attach disk [datastore1] bc13bbe2-6e02-4d98-9e50-94f772d89ac0/bc13bbe2-6e02-4d98-9e50-94f772d89ac0.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.588385] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-317a1441-6cd6-47d5-893c-3f7fe22cc6ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.615295] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f009e8-4b70-2906-e14a-1bc1b044f10e, 'name': SearchDatastore_Task, 'duration_secs': 0.026657} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.621124] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 695.621124] env[66641]: value = "task-5145745" [ 695.621124] env[66641]: _type = "Task" [ 695.621124] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.621124] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9187f63c-dd10-4a38-a323-3ebb80ea140d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.633393] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.636987] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 695.636987] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52524bf4-3602-ddfe-5fd0-16ff7b0472e9" [ 695.636987] env[66641]: _type = "Task" [ 695.636987] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.643033] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145745, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.655096] env[66641]: DEBUG oslo_concurrency.lockutils [req-599f9011-8357-408c-ad92-d914e21840d8 req-382a1de8-c55b-4625-b60d-3c98c4452556 service nova] Releasing lock "refresh_cache-bc13bbe2-6e02-4d98-9e50-94f772d89ac0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.655096] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52524bf4-3602-ddfe-5fd0-16ff7b0472e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.754740] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145744, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.775174] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 695.775608] env[66641]: WARNING openstack [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 695.812129] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Releasing lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 695.812541] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Instance network_info: |[{"id": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "address": "fa:16:3e:6f:e9:79", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.49", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932f4137-a3", "ovs_interfaceid": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 695.813074] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:e9:79', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '932f4137-a36a-42bf-9b2f-ebfeaa078c3c', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 695.820751] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Creating folder: Project (95981b554a7f41e189993a89ea395b43). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.821054] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5c4caf1a-ad8f-4989-ba9a-7e16d6a32a7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.834283] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Created folder: Project (95981b554a7f41e189993a89ea395b43) in parent group-v1000566. [ 695.834473] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Creating folder: Instances. Parent ref: group-v1000628. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 695.834652] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc4e2031-d283-4d69-95f3-d002bc66aa41 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.847144] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Created folder: Instances in parent group-v1000628. [ 695.847489] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 695.849292] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 695.849292] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e5202e2e-4791-4b6e-9af2-af67b14f90e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.877026] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 695.877026] env[66641]: value = "task-5145748" [ 695.877026] env[66641]: _type = "Task" [ 695.877026] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.884272] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145748, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.050593] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.051016] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.064428] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Updated VIF entry in instance network info cache for port f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 696.065110] env[66641]: DEBUG nova.network.neutron [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Updating instance_info_cache with network_info: [{"id": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "address": "fa:16:3e:84:b1:ad", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.95", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf4cb47fb-c2", "ovs_interfaceid": "f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 696.068545] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 696.069900] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-36969392-fd43-49ea-a242-28f70be32f3f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.080871] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 696.080871] env[66641]: value = "task-5145749" [ 696.080871] env[66641]: _type = "Task" [ 696.080871] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.096458] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145749, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.128140] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.138382] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145745, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.144065] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.144544] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.150501] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.150634] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.150787] env[66641]: DEBUG nova.network.neutron [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 696.150964] env[66641]: DEBUG nova.objects.instance [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lazy-loading 'info_cache' on Instance uuid dfa8c73b-db57-42a9-a9a4-cf812f5b2949 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 696.173075] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52524bf4-3602-ddfe-5fd0-16ff7b0472e9, 'name': SearchDatastore_Task, 'duration_secs': 0.021327} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.173528] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.176483] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] ac5bb5a2-e630-42f0-be0c-24f52be81367/ac5bb5a2-e630-42f0-be0c-24f52be81367.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 696.176820] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a8ea66e-5722-4801-9bb9-2135b71b15a3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.189586] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 696.189586] env[66641]: value = "task-5145750" [ 696.189586] env[66641]: _type = "Task" [ 696.189586] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.204684] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145750, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.253913] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145744, 'name': CreateVM_Task, 'duration_secs': 0.644126} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.253913] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 696.254627] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.255204] env[66641]: WARNING openstack [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.261307] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.261307] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.261822] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 696.262280] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbb55633-7ff8-4d65-a100-a4fb0ec86d72 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.269336] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 696.269336] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525283ce-ee37-12ce-00c0-acbc33e16acb" [ 696.269336] env[66641]: _type = "Task" [ 696.269336] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.280614] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525283ce-ee37-12ce-00c0-acbc33e16acb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.385840] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145748, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.568281] env[66641]: DEBUG oslo_concurrency.lockutils [req-0a2f5396-8846-4ed2-84c1-e2ba4520e52c req-71e33672-682e-48f5-b631-2a09917b0919 service nova] Releasing lock "refresh_cache-22899269-28fc-446e-9533-1f6862a9e8db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.593793] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145749, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.626161] env[66641]: DEBUG oslo_vmware.api [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145741, 'name': PowerOnVM_Task, 'duration_secs': 1.293511} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.626456] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 696.626763] env[66641]: DEBUG nova.compute.manager [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 696.627559] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c4e292-9ff0-4bcb-bf55-7316872b7c1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.639991] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145745, 'name': ReconfigVM_Task, 'duration_secs': 0.561621} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.641974] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Reconfigured VM instance instance-00000013 to attach disk [datastore1] bc13bbe2-6e02-4d98-9e50-94f772d89ac0/bc13bbe2-6e02-4d98-9e50-94f772d89ac0.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 696.645419] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45cd647c-5d09-480d-b86a-e8fc8356bf1f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.654561] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 696.654561] env[66641]: value = "task-5145751" [ 696.654561] env[66641]: _type = "Task" [ 696.654561] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.667264] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145751, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.701760] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145750, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.782395] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525283ce-ee37-12ce-00c0-acbc33e16acb, 'name': SearchDatastore_Task, 'duration_secs': 0.015889} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.782724] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 696.782967] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 696.783342] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.783493] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.783661] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 696.783944] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1f000be0-b313-4327-95e3-8168089d5859 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.800607] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 696.800807] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 696.801801] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-132cbd45-96ad-418b-9528-2ab99c2577c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.809684] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 696.809684] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb9c85-a7f2-2a5b-950a-6e3d4c11a5c6" [ 696.809684] env[66641]: _type = "Task" [ 696.809684] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.819297] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb9c85-a7f2-2a5b-950a-6e3d4c11a5c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.884569] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145748, 'name': CreateVM_Task, 'duration_secs': 0.869525} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 696.884799] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 696.885693] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 696.886189] env[66641]: WARNING openstack [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 696.892078] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.892364] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 696.892747] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 696.893081] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4c9963bc-b30e-4cec-a10f-1d75ad00f956 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.899969] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 696.899969] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529bdec4-358a-0dbc-cc06-56efeb98a473" [ 696.899969] env[66641]: _type = "Task" [ 696.899969] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.914358] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529bdec4-358a-0dbc-cc06-56efeb98a473, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.954826] env[66641]: DEBUG nova.compute.manager [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 696.955884] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48a9f154-6a2f-4223-950a-ee33f9f7d989 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.093488] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145749, 'name': CreateSnapshot_Task, 'duration_secs': 0.777356} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.094201] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 697.094673] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a95e43-c113-4b1d-9837-d0f592b2e0dd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.159718] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 697.160015] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 697.160254] env[66641]: DEBUG nova.objects.instance [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 697.169480] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.171217] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.189345] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145751, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.200371] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145750, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719711} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.201462] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] ac5bb5a2-e630-42f0-be0c-24f52be81367/ac5bb5a2-e630-42f0-be0c-24f52be81367.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 697.201462] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 697.201462] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0da4a9fc-56da-4c18-8f96-328c736c4621 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.209850] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 697.209850] env[66641]: value = "task-5145752" [ 697.209850] env[66641]: _type = "Task" [ 697.209850] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.224761] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145752, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.322512] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb9c85-a7f2-2a5b-950a-6e3d4c11a5c6, 'name': SearchDatastore_Task, 'duration_secs': 0.05614} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.323844] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26da47fb-8d4b-4e0d-8d09-c88eee54ce1c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.330486] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 697.330486] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fe9007-d62a-4186-f26b-894260536841" [ 697.330486] env[66641]: _type = "Task" [ 697.330486] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.340130] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fe9007-d62a-4186-f26b-894260536841, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.416171] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529bdec4-358a-0dbc-cc06-56efeb98a473, 'name': SearchDatastore_Task, 'duration_secs': 0.015437} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.416171] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.416171] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 697.416171] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.469410] env[66641]: INFO nova.compute.manager [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] instance snapshotting [ 697.470376] env[66641]: WARNING nova.compute.manager [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 697.477432] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3670c6e4-19b0-40c7-bc39-79c0d94e99af {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.499695] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1957acb-52d2-49ae-a9f5-0c885b696234 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.617569] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 697.617965] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8bed4a45-fd4f-488e-a3f8-ecb3dad97f25 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.630981] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 697.630981] env[66641]: value = "task-5145753" [ 697.630981] env[66641]: _type = "Task" [ 697.630981] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.638267] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145753, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.670909] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145751, 'name': Rename_Task, 'duration_secs': 0.541609} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.670909] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 697.671365] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d666e37e-0be6-4f03-8fa4-2f517a36144a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.690383] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 697.690383] env[66641]: value = "task-5145754" [ 697.690383] env[66641]: _type = "Task" [ 697.690383] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.710089] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145754, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.722521] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145752, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310962} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.722979] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 697.724297] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e26018a-ddb8-44fc-bc04-1d8fd84fe0ec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.760746] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Reconfiguring VM instance instance-00000014 to attach disk [datastore1] ac5bb5a2-e630-42f0-be0c-24f52be81367/ac5bb5a2-e630-42f0-be0c-24f52be81367.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 697.760859] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-13e4e134-87c8-4c13-926b-ec252a7fcf97 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.787572] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 697.787572] env[66641]: value = "task-5145755" [ 697.787572] env[66641]: _type = "Task" [ 697.787572] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.797753] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.846997] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fe9007-d62a-4186-f26b-894260536841, 'name': SearchDatastore_Task, 'duration_secs': 0.046147} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.846997] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 697.846997] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] a2d567c9-09af-44e1-aad2-c8f394a09714/a2d567c9-09af-44e1-aad2-c8f394a09714.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 697.847475] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 697.847518] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 697.848276] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9916fd19-9cca-4da4-95d7-e741050de0f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.849955] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e827cbf9-5d75-45bb-bde1-24390bc37777 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.853402] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.853775] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 697.867321] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 697.867321] env[66641]: value = "task-5145756" [ 697.867321] env[66641]: _type = "Task" [ 697.867321] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.873552] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 697.873810] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 697.877804] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f60f1776-054d-4a71-b1f1-2b0f3941ab0d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.885830] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.889531] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 697.889531] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52410a5a-ca98-2aaf-be0a-0433edb9bed5" [ 697.889531] env[66641]: _type = "Task" [ 697.889531] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.900863] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52410a5a-ca98-2aaf-be0a-0433edb9bed5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.977596] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 697.978466] env[66641]: WARNING openstack [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.018749] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 698.019914] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f89064e5-ac80-4633-ab69-f13e2f835822 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.030761] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 698.030761] env[66641]: value = "task-5145757" [ 698.030761] env[66641]: _type = "Task" [ 698.030761] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.046130] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145757, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.096872] env[66641]: DEBUG nova.compute.manager [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Received event network-vif-plugged-932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 698.097253] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Acquiring lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.097457] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.098231] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.098500] env[66641]: DEBUG nova.compute.manager [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] No waiting events found dispatching network-vif-plugged-932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 698.098677] env[66641]: WARNING nova.compute.manager [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Received unexpected event network-vif-plugged-932f4137-a36a-42bf-9b2f-ebfeaa078c3c for instance with vm_state building and task_state spawning. [ 698.098830] env[66641]: DEBUG nova.compute.manager [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Received event network-changed-932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 698.098979] env[66641]: DEBUG nova.compute.manager [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Refreshing instance network info cache due to event network-changed-932f4137-a36a-42bf-9b2f-ebfeaa078c3c. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 698.099181] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Acquiring lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.099312] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Acquired lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 698.099525] env[66641]: DEBUG nova.network.neutron [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Refreshing network info cache for port 932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 698.139881] env[66641]: DEBUG nova.network.neutron [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [{"id": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "address": "fa:16:3e:8a:b6:99", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.160", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5ebec271-9d", "ovs_interfaceid": "5ebec271-9d57-41f2-ba8c-59edb82c50db", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 698.149401] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145753, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.192976] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e269714c-e521-46a0-9a0b-8f9a6008848a tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.033s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 698.211090] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145754, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.299533] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145755, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.380642] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145756, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.403860] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52410a5a-ca98-2aaf-be0a-0433edb9bed5, 'name': SearchDatastore_Task, 'duration_secs': 0.024811} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.404674] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-900a0765-48f0-4088-92ff-aca48eec700b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.412509] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 698.412509] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528f1f83-ed3e-d6ee-64f2-04c7e258ab38" [ 698.412509] env[66641]: _type = "Task" [ 698.412509] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.424809] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528f1f83-ed3e-d6ee-64f2-04c7e258ab38, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.549941] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145757, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.607033] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.607656] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.643060] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-dfa8c73b-db57-42a9-a9a4-cf812f5b2949" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.643445] env[66641]: DEBUG nova.objects.instance [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lazy-loading 'migration_context' on Instance uuid dfa8c73b-db57-42a9-a9a4-cf812f5b2949 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 698.646706] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145753, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.709851] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145754, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.772023] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.772023] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.799085] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145755, 'name': ReconfigVM_Task, 'duration_secs': 0.735577} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.799526] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Reconfigured VM instance instance-00000014 to attach disk [datastore1] ac5bb5a2-e630-42f0-be0c-24f52be81367/ac5bb5a2-e630-42f0-be0c-24f52be81367.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 698.800402] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a5d34db-2a41-4e10-96e9-8f09f78189da {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.808797] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 698.808797] env[66641]: value = "task-5145758" [ 698.808797] env[66641]: _type = "Task" [ 698.808797] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.817779] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 698.818038] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 698.825692] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145758, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.877191] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 698.877581] env[66641]: WARNING openstack [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 698.899165] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692093} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.899445] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] a2d567c9-09af-44e1-aad2-c8f394a09714/a2d567c9-09af-44e1-aad2-c8f394a09714.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 698.899653] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 698.899925] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-682385c4-56a0-492d-8270-38d03aa74c9a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.908079] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 698.908079] env[66641]: value = "task-5145759" [ 698.908079] env[66641]: _type = "Task" [ 698.908079] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.919799] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.926757] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528f1f83-ed3e-d6ee-64f2-04c7e258ab38, 'name': SearchDatastore_Task, 'duration_secs': 0.074123} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.927083] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 698.927412] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] eeb2556a-4e0c-43d7-83bd-942be5d5fdd2/eeb2556a-4e0c-43d7-83bd-942be5d5fdd2.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 698.927698] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-138e9d95-15ba-4497-a380-23be5e5e7805 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.937529] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 698.937529] env[66641]: value = "task-5145760" [ 698.937529] env[66641]: _type = "Task" [ 698.937529] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.956594] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145760, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.996575] env[66641]: DEBUG nova.network.neutron [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Updated VIF entry in instance network info cache for port 932f4137-a36a-42bf-9b2f-ebfeaa078c3c. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 698.996979] env[66641]: DEBUG nova.network.neutron [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Updating instance_info_cache with network_info: [{"id": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "address": "fa:16:3e:6f:e9:79", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.49", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap932f4137-a3", "ovs_interfaceid": "932f4137-a36a-42bf-9b2f-ebfeaa078c3c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 699.044445] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145757, 'name': CreateSnapshot_Task, 'duration_secs': 0.617962} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.044764] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 699.045965] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0604a851-df7b-4c29-b5e2-96382350f003 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.143278] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145753, 'name': CloneVM_Task, 'duration_secs': 1.469233} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.143670] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Created linked-clone VM from snapshot [ 699.145044] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d0be2cb-957e-46e1-9e7d-e0d5b538963e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.149782] env[66641]: DEBUG nova.objects.base [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=66641) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 699.151854] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b29205-347d-4cac-a8aa-d6cdf372965d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.162021] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Uploading image 1b88364d-239d-4b24-b05f-cf0ee4b256b3 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 699.180176] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34203ae4-ff2d-4d38-8494-936b83685725 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.187492] env[66641]: DEBUG oslo_vmware.api [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 699.187492] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3ef2c-df12-3879-8d85-5233ea2eba8f" [ 699.187492] env[66641]: _type = "Task" [ 699.187492] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.197030] env[66641]: DEBUG oslo_vmware.api [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3ef2c-df12-3879-8d85-5233ea2eba8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.207976] env[66641]: DEBUG oslo_vmware.api [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145754, 'name': PowerOnVM_Task, 'duration_secs': 1.07031} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.210530] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 699.210530] env[66641]: value = "vm-1000632" [ 699.210530] env[66641]: _type = "VirtualMachine" [ 699.210530] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 699.210981] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 699.211450] env[66641]: INFO nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Took 15.21 seconds to spawn the instance on the hypervisor. [ 699.211773] env[66641]: DEBUG nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 699.211931] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-4f3a3deb-1bbe-41fa-9130-5dfa7275dc56 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.214185] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40415db7-b713-40b8-acfc-079b393e735f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.226190] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lease: (returnval){ [ 699.226190] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523382a9-9ca5-d039-9e8b-c37b7d6bc053" [ 699.226190] env[66641]: _type = "HttpNfcLease" [ 699.226190] env[66641]: } obtained for exporting VM: (result){ [ 699.226190] env[66641]: value = "vm-1000632" [ 699.226190] env[66641]: _type = "VirtualMachine" [ 699.226190] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 699.226768] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the lease: (returnval){ [ 699.226768] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523382a9-9ca5-d039-9e8b-c37b7d6bc053" [ 699.226768] env[66641]: _type = "HttpNfcLease" [ 699.226768] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 699.234596] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 699.234596] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523382a9-9ca5-d039-9e8b-c37b7d6bc053" [ 699.234596] env[66641]: _type = "HttpNfcLease" [ 699.234596] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 699.285180] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.285784] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.286083] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.286554] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.287394] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.290205] env[66641]: INFO nova.compute.manager [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Terminating instance [ 699.322803] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 699.328026] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145758, 'name': Rename_Task, 'duration_secs': 0.223348} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.328026] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 699.328026] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-92ca395d-eab5-44f1-8d1d-256b63f044e7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.336112] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 699.336112] env[66641]: value = "task-5145762" [ 699.336112] env[66641]: _type = "Task" [ 699.336112] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.348098] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145762, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.424555] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104801} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.427021] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Received event network-vif-plugged-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 699.427021] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquiring lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.427021] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.427021] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.427021] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] No waiting events found dispatching network-vif-plugged-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 699.427434] env[66641]: WARNING nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Received unexpected event network-vif-plugged-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 for instance with vm_state building and task_state spawning. [ 699.427434] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Received event network-changed-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 699.427434] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Refreshing instance network info cache due to event network-changed-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 699.428209] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquiring lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.428209] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquired lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 699.428209] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Refreshing network info cache for port 33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 699.429788] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 699.431664] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25c60375-d301-42f4-914f-bcd70064fad6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.473850] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Reconfiguring VM instance instance-00000015 to attach disk [datastore1] a2d567c9-09af-44e1-aad2-c8f394a09714/a2d567c9-09af-44e1-aad2-c8f394a09714.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 699.474932] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bd59e40c-a8a6-4a7c-830d-dec03715e0dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.500450] env[66641]: DEBUG oslo_concurrency.lockutils [req-1a9770db-981a-4758-84b5-69406f5675cd req-244ab3e2-42b9-4dcc-9293-b012c3358dad service nova] Releasing lock "refresh_cache-eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 699.500990] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145760, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.504320] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 699.504320] env[66641]: value = "task-5145763" [ 699.504320] env[66641]: _type = "Task" [ 699.504320] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.515071] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145763, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.567448] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 699.567761] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-ac2f7403-db14-4af2-879f-a2db0e548a83 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.582443] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 699.582443] env[66641]: value = "task-5145764" [ 699.582443] env[66641]: _type = "Task" [ 699.582443] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.599980] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.706421] env[66641]: DEBUG oslo_vmware.api [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3ef2c-df12-3879-8d85-5233ea2eba8f, 'name': SearchDatastore_Task, 'duration_secs': 0.011297} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.706777] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.707041] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.743639] env[66641]: INFO nova.compute.manager [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Took 32.99 seconds to build instance. [ 699.753979] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 699.753979] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523382a9-9ca5-d039-9e8b-c37b7d6bc053" [ 699.753979] env[66641]: _type = "HttpNfcLease" [ 699.753979] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 699.753979] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 699.753979] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523382a9-9ca5-d039-9e8b-c37b7d6bc053" [ 699.753979] env[66641]: _type = "HttpNfcLease" [ 699.753979] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 699.753979] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66bd5843-7433-4599-935d-ccbd31b692c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.769597] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 699.769997] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 699.840223] env[66641]: DEBUG nova.compute.manager [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 699.840594] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 699.845557] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d6bf41-4448-400e-a3bd-3bd4bdae9bf5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.864794] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145762, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.867804] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 699.868105] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bb5bd4b2-3735-4ee3-b83b-2fc2df4971d3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.877819] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 699.877819] env[66641]: value = "task-5145765" [ 699.877819] env[66641]: _type = "Task" [ 699.877819] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.889086] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.889736] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.889964] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.890621] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.890621] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.891081] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 699.893932] env[66641]: INFO nova.compute.manager [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Terminating instance [ 699.900264] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145765, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.932828] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 699.933231] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 699.957257] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3f4dfdca-84c5-4b08-ac3e-2aa5cc8482a0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.969237] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "8235441c-c046-49da-a5a5-92bd85e17982" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 699.969724] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 699.980583] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 699.982023] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145760, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.818325} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.983385] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1350ce-101d-49d6-967a-b14cfce3ffe0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.988511] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] eeb2556a-4e0c-43d7-83bd-942be5d5fdd2/eeb2556a-4e0c-43d7-83bd-942be5d5fdd2.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 699.990034] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 699.990204] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bb482d6c-284a-4eec-b3c6-8b562f53b92e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.004253] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 700.004624] env[66641]: ERROR oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk due to incomplete transfer.: nova.exception.ImageNotFound: Image 1b88364d-239d-4b24-b05f-cf0ee4b256b3 could not be found. [ 700.008268] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a0d80b37-e098-49a5-a013-36f6d422beb8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.010539] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 700.010539] env[66641]: value = "task-5145766" [ 700.010539] env[66641]: _type = "Task" [ 700.010539] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.026926] env[66641]: DEBUG oslo_vmware.rw_handles [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/521fc504-c7f8-8c5f-4eef-b8933d5a7794/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 700.029070] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 700.029070] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-22f4645d-87da-43ab-b79d-e2a48fe2771f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.041525] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145763, 'name': ReconfigVM_Task, 'duration_secs': 0.498387} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.041911] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.043149] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Reconfigured VM instance instance-00000015 to attach disk [datastore1] a2d567c9-09af-44e1-aad2-c8f394a09714/a2d567c9-09af-44e1-aad2-c8f394a09714.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 700.047016] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5e63e470-50b3-4226-9cb7-ac7a5e671c3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.050631] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 700.050631] env[66641]: value = "task-5145767" [ 700.050631] env[66641]: _type = "Task" [ 700.050631] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.060615] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 700.060615] env[66641]: value = "task-5145768" [ 700.060615] env[66641]: _type = "Task" [ 700.060615] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.069298] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145767, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.077650] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145768, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.095061] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.105805] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "d03f39a4-532b-439a-9055-19fc1e769fff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 700.105971] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 700.250829] env[66641]: DEBUG oslo_concurrency.lockutils [None req-619e5575-641c-44fe-b5e0-ff70dfaa66be tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.518s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 700.253185] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.253560] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.391777] env[66641]: DEBUG oslo_vmware.api [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145762, 'name': PowerOnVM_Task, 'duration_secs': 0.720729} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.391777] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 700.391777] env[66641]: INFO nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Took 13.97 seconds to spawn the instance on the hypervisor. [ 700.391777] env[66641]: DEBUG nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 700.391777] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fe9b31-170e-47ff-96df-d3e002a3558c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.395514] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145765, 'name': PowerOffVM_Task, 'duration_secs': 0.38671} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.395514] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 700.395514] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 700.395514] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d4cad83c-1fd2-4ee0-87c9-531dd8971832 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.405769] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "refresh_cache-5963a8ba-9d2c-42c3-a5da-25c29bf9e763" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 700.405769] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquired lock "refresh_cache-5963a8ba-9d2c-42c3-a5da-25c29bf9e763" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 700.405769] env[66641]: DEBUG nova.network.neutron [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 700.486470] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 700.486470] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 700.486470] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleting the datastore file [datastore2] 434e66b1-5ea7-43d2-9b41-bdeda2a312d7 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 700.489323] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-527a6269-c72a-4a4c-a84a-3ca05bd70381 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.491041] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 700.501434] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 700.501434] env[66641]: value = "task-5145770" [ 700.501434] env[66641]: _type = "Task" [ 700.501434] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.515209] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145770, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.539193] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.161549} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.539193] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.539193] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5dec4b0-8b81-4260-a192-3781de5e887b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.561826] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Reconfiguring VM instance instance-00000016 to attach disk [datastore1] eeb2556a-4e0c-43d7-83bd-942be5d5fdd2/eeb2556a-4e0c-43d7-83bd-942be5d5fdd2.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 700.568703] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bba4653-ecaa-498f-b2ed-f113b67cef06 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.599531] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145767, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.601826] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 700.601826] env[66641]: value = "task-5145771" [ 700.601826] env[66641]: _type = "Task" [ 700.601826] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.607075] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145768, 'name': Rename_Task, 'duration_secs': 0.283751} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.612955] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 700.613397] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 700.617246] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.622821] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6187eaf3-2b8a-4091-a5bc-1107f55f02be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.636431] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145771, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.637020] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 700.637020] env[66641]: value = "task-5145772" [ 700.637020] env[66641]: _type = "Task" [ 700.637020] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.647691] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145772, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.792689] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a139474-d660-444f-9b27-0917d0eb4e10 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.806592] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ca5e963-ae86-4164-ad52-4969b3f1328c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.843512] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948bf1a2-80e2-4bb8-9f0c-09cb3e1b23c7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.852685] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eadf6b-4232-459f-8359-1d5452fa7c86 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.868603] env[66641]: DEBUG nova.compute.provider_tree [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.892386] env[66641]: INFO nova.compute.manager [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Took 24.42 seconds to build instance. [ 700.940169] env[66641]: WARNING openstack [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.940767] env[66641]: WARNING openstack [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 700.948268] env[66641]: DEBUG nova.network.neutron [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 700.981082] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 700.981774] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.016408] env[66641]: DEBUG oslo_vmware.api [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145770, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.270994} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.016719] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 701.016921] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 701.017113] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.017348] env[66641]: INFO nova.compute.manager [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Took 1.18 seconds to destroy the instance on the hypervisor. [ 701.017680] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 701.017945] env[66641]: DEBUG nova.compute.manager [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 701.018089] env[66641]: DEBUG nova.network.neutron [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 701.018710] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.019084] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.026676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.065429] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145767, 'name': Destroy_Task, 'duration_secs': 0.668494} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.065681] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Destroyed the VM [ 701.066014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 701.066276] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-3690721c-8be0-42d6-8234-f0722fabccbc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.074694] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 701.074694] env[66641]: value = "task-5145773" [ 701.074694] env[66641]: _type = "Task" [ 701.074694] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.084771] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145773, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.100592] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.118408] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145771, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.158219] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145772, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.159650] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 701.359730] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 701.360273] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 701.372178] env[66641]: DEBUG nova.scheduler.client.report [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 701.394612] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a0ac8387-7eda-4f6a-8f14-2025eabc8f3b tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.933s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 701.589641] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145773, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.604235] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.616566] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145771, 'name': ReconfigVM_Task, 'duration_secs': 0.866207} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.616858] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Reconfigured VM instance instance-00000016 to attach disk [datastore1] eeb2556a-4e0c-43d7-83bd-942be5d5fdd2/eeb2556a-4e0c-43d7-83bd-942be5d5fdd2.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 701.617581] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e53fafdd-f8a6-4eb2-8b79-f4521c3bd014 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.626528] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 701.626528] env[66641]: value = "task-5145774" [ 701.626528] env[66641]: _type = "Task" [ 701.626528] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.636902] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145774, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.647556] env[66641]: DEBUG oslo_vmware.api [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145772, 'name': PowerOnVM_Task, 'duration_secs': 1.011457} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.647860] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 701.648147] env[66641]: INFO nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Took 12.79 seconds to spawn the instance on the hypervisor. [ 701.648374] env[66641]: DEBUG nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 701.649230] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2883150e-e403-42a0-b2a1-e8616cb25608 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.086368] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145773, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.105220] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145764, 'name': CloneVM_Task, 'duration_secs': 2.163816} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.105220] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Created linked-clone VM from snapshot [ 702.106049] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb425f41-aadc-41dc-a3e6-157083f5588f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.115557] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Uploading image b6f3d72e-8bf4-4ee3-a9f9-3195a8e0534e {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 702.140546] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145774, 'name': Rename_Task, 'duration_secs': 0.296163} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.140902] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 702.141225] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52b3ffe1-511e-4854-8e84-882e53da1b91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.150772] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 702.150772] env[66641]: value = "task-5145775" [ 702.150772] env[66641]: _type = "Task" [ 702.150772] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.153582] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 702.153582] env[66641]: value = "vm-1000634" [ 702.153582] env[66641]: _type = "VirtualMachine" [ 702.153582] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 702.154149] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a57a50fe-2a20-4703-94b5-147fa3c560f7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.171907] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145775, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.176155] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lease: (returnval){ [ 702.176155] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e51a92-13e3-4bed-a7b4-3ba269f07449" [ 702.176155] env[66641]: _type = "HttpNfcLease" [ 702.176155] env[66641]: } obtained for exporting VM: (result){ [ 702.176155] env[66641]: value = "vm-1000634" [ 702.176155] env[66641]: _type = "VirtualMachine" [ 702.176155] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 702.176155] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the lease: (returnval){ [ 702.176155] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e51a92-13e3-4bed-a7b4-3ba269f07449" [ 702.176155] env[66641]: _type = "HttpNfcLease" [ 702.176155] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 702.180371] env[66641]: INFO nova.compute.manager [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Took 20.78 seconds to build instance. [ 702.185751] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 702.185751] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e51a92-13e3-4bed-a7b4-3ba269f07449" [ 702.185751] env[66641]: _type = "HttpNfcLease" [ 702.185751] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 702.355132] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Updated VIF entry in instance network info cache for port 33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 702.355541] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Updating instance_info_cache with network_info: [{"id": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "address": "fa:16:3e:fc:ab:bb", "network": {"id": "114a8269-be9e-4964-abfe-24aee4c612cd", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-1298879401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ef1e42e42cd492d8dce962e90011df8", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7654928b-7afe-42e3-a18d-68ecc775cefe", "external-id": "cl2-zone-807", "segmentation_id": 807, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33cb5bcf-2e", "ovs_interfaceid": "33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.385912] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.679s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.392419] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.503s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.396477] env[66641]: INFO nova.compute.claims [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.589899] env[66641]: DEBUG oslo_vmware.api [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145773, 'name': RemoveSnapshot_Task, 'duration_secs': 1.331142} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.590887] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 702.609089] env[66641]: DEBUG nova.network.neutron [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 702.668572] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145775, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.682833] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9a3ac37a-cee4-4cea-b5d5-f2f614c4f6ef tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.304s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.690208] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 702.690208] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e51a92-13e3-4bed-a7b4-3ba269f07449" [ 702.690208] env[66641]: _type = "HttpNfcLease" [ 702.690208] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 702.690208] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 702.690208] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e51a92-13e3-4bed-a7b4-3ba269f07449" [ 702.690208] env[66641]: _type = "HttpNfcLease" [ 702.690208] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 702.690902] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757b5d42-77a7-4e92-93e5-58ec6d5ae6af {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.704635] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 702.704923] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 702.862450] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Releasing lock "refresh_cache-ac5bb5a2-e630-42f0-be0c-24f52be81367" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 702.862898] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Received event network-vif-plugged-d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 702.863112] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquiring lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 702.863615] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 702.863784] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 702.863956] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] No waiting events found dispatching network-vif-plugged-d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 702.864376] env[66641]: WARNING nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Received unexpected event network-vif-plugged-d66320d7-4884-46e8-83ab-a8a6fd0b09a7 for instance with vm_state building and task_state spawning. [ 702.864376] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Received event network-changed-d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 702.864475] env[66641]: DEBUG nova.compute.manager [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Refreshing instance network info cache due to event network-changed-d66320d7-4884-46e8-83ab-a8a6fd0b09a7. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 702.864609] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquiring lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.864739] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Acquired lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 702.864889] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Refreshing network info cache for port d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 702.866686] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-496b5c5e-f563-42e3-805c-2df3b1f3521b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.993560] env[66641]: INFO nova.scheduler.client.report [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted allocation for migration 1ed74871-41c7-4c89-bbf4-b55c237571e3 [ 703.083915] env[66641]: DEBUG nova.network.neutron [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 703.096642] env[66641]: WARNING nova.compute.manager [None req-88a2c807-568c-411f-b539-d5e6c3e018f2 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Image not found during snapshot: nova.exception.ImageNotFound: Image 1b88364d-239d-4b24-b05f-cf0ee4b256b3 could not be found. [ 703.114640] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Releasing lock "refresh_cache-5963a8ba-9d2c-42c3-a5da-25c29bf9e763" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 703.115032] env[66641]: DEBUG nova.compute.manager [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 703.117380] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.118255] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c67910d-ec47-4d0e-b6eb-4ca3e32f6d65 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.127419] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 703.127682] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-12d652f1-46fc-4783-9e20-a9696312afee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.141717] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 703.141717] env[66641]: value = "task-5145777" [ 703.141717] env[66641]: _type = "Task" [ 703.141717] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.154688] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145777, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.167908] env[66641]: DEBUG oslo_vmware.api [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145775, 'name': PowerOnVM_Task, 'duration_secs': 0.806778} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.168214] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 703.168428] env[66641]: INFO nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Took 11.39 seconds to spawn the instance on the hypervisor. [ 703.168719] env[66641]: DEBUG nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 703.169573] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c9a4467-9cd5-44e6-8d16-9a1b083b43ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.373132] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.373132] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.503997] env[66641]: DEBUG oslo_concurrency.lockutils [None req-987eecba-fbfa-47dd-ab9b-8dc3591ee21b tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.970s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 703.588821] env[66641]: INFO nova.compute.manager [-] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Took 2.57 seconds to deallocate network for instance. [ 703.659618] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145777, 'name': PowerOffVM_Task, 'duration_secs': 0.163704} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 703.666729] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 703.668086] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 703.668086] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7cb6ae2b-7efa-4c1a-9402-23200fc5d514 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.701269] env[66641]: INFO nova.compute.manager [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Took 21.98 seconds to build instance. [ 703.704955] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 703.705407] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 703.705533] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Deleting the datastore file [datastore1] 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 703.707146] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80de9997-0906-40f3-8a22-153c7d1a54d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.715265] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for the task: (returnval){ [ 703.715265] env[66641]: value = "task-5145779" [ 703.715265] env[66641]: _type = "Task" [ 703.715265] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.742106] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.885432] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afee629f-50f5-4b64-8de8-63fa281e3fa2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.903084] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d867762-298a-4a5d-9f65-47177d02d93d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.950653] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2f6219-12e4-480e-8d19-cc55206d09aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.970078] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92744ce3-519f-4d72-b859-044f8eaac81a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.976270] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 703.976669] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 703.998754] env[66641]: DEBUG nova.compute.provider_tree [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.056636] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.058020] env[66641]: WARNING openstack [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.096141] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 704.208067] env[66641]: DEBUG oslo_concurrency.lockutils [None req-adee15f5-e4de-41db-9452-e07eea42c550 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.497s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 704.227071] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Updated VIF entry in instance network info cache for port d66320d7-4884-46e8-83ab-a8a6fd0b09a7. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 704.227192] env[66641]: DEBUG nova.network.neutron [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Updating instance_info_cache with network_info: [{"id": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "address": "fa:16:3e:13:e7:e4", "network": {"id": "18515ee6-2926-44bf-b2bc-d5101fd6de51", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1940026041-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "42e99a1d4fdb478d825190d12850ad89", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b356db78-99c7-4464-822c-fc7e193f7878", "external-id": "nsx-vlan-transportzone-231", "segmentation_id": 231, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd66320d7-48", "ovs_interfaceid": "d66320d7-4884-46e8-83ab-a8a6fd0b09a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.235958] env[66641]: DEBUG oslo_vmware.api [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Task: {'id': task-5145779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246475} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.236501] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 704.236501] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 704.236692] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.236857] env[66641]: INFO nova.compute.manager [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Took 1.12 seconds to destroy the instance on the hypervisor. [ 704.237476] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 704.237737] env[66641]: DEBUG nova.compute.manager [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 704.237856] env[66641]: DEBUG nova.network.neutron [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 704.241396] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.241396] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.295335] env[66641]: DEBUG nova.network.neutron [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 704.301029] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 704.301029] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 704.502926] env[66641]: DEBUG nova.scheduler.client.report [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 704.731265] env[66641]: DEBUG oslo_concurrency.lockutils [req-4865b647-91bb-4a0b-ba32-a0e78e9c8d6a req-c9551800-9f17-489f-9f82-219b8dbd5867 service nova] Releasing lock "refresh_cache-a2d567c9-09af-44e1-aad2-c8f394a09714" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 704.816849] env[66641]: DEBUG nova.network.neutron [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 704.818839] env[66641]: DEBUG nova.compute.manager [req-2e60a066-ea7d-47ae-95a7-f3deed605c53 req-91642c28-a78b-40d0-81e5-c41610ff5af2 service nova] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Received event network-vif-deleted-2fd71b07-2d9b-426a-a09a-715cea2538e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 705.012775] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.618s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.012775] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 705.013775] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.987s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.015951] env[66641]: INFO nova.compute.claims [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.106063] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "22899269-28fc-446e-9533-1f6862a9e8db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.106348] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.106575] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "22899269-28fc-446e-9533-1f6862a9e8db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.106759] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.106920] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.112096] env[66641]: INFO nova.compute.manager [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Terminating instance [ 705.324101] env[66641]: INFO nova.compute.manager [-] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Took 1.09 seconds to deallocate network for instance. [ 705.522051] env[66641]: DEBUG nova.compute.utils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 705.525592] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 705.525842] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 705.526187] env[66641]: WARNING neutronclient.v2_0.client [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.526564] env[66641]: WARNING neutronclient.v2_0.client [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 705.529518] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 705.531282] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 705.617129] env[66641]: DEBUG nova.compute.manager [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 705.617355] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 705.619083] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77e165ed-d8e1-468a-875a-1077b924ec68 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.634221] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 705.634221] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-43dc66eb-f421-4dab-9225-fd751d6b4058 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.645249] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 705.645249] env[66641]: value = "task-5145780" [ 705.645249] env[66641]: _type = "Task" [ 705.645249] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.658545] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.832478] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.890189] env[66641]: DEBUG nova.policy [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba7ff9aa8f34a7da6344773cea86836', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2c8481015524aee95a933f61082faec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 705.975613] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.978510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.978510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 705.978510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 705.978510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 705.982365] env[66641]: INFO nova.compute.manager [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Terminating instance [ 706.029024] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 706.165395] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145780, 'name': PowerOffVM_Task, 'duration_secs': 0.31913} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.165395] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 706.165538] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 706.165831] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d901caf-b001-4d06-bf58-526901effcfc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.385442] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9737d066-04f6-45d9-ba0f-5b573ad7f85b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.403849] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-465de35d-25a1-4cc4-a66a-4348c45e2426 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.450304] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7a74c5-e82e-4341-b110-553c5e05ce63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.461031] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674df546-1f0f-48aa-9bae-bc20c1eb1723 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.480422] env[66641]: DEBUG nova.compute.provider_tree [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.489591] env[66641]: DEBUG nova.compute.manager [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 706.489827] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 706.491189] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea41dccf-2a3f-4d8b-81d6-8f2b5aa5e355 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.501587] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 706.502646] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 706.502646] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Deleting the datastore file [datastore2] 22899269-28fc-446e-9533-1f6862a9e8db {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 706.505865] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-544fabf3-1af6-4f5b-8232-5f8e11e17dbc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.513042] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 706.513042] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f27f7146-2cf8-4848-b128-ba281bd3cc01 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.516920] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for the task: (returnval){ [ 706.516920] env[66641]: value = "task-5145782" [ 706.516920] env[66641]: _type = "Task" [ 706.516920] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.525142] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 706.525142] env[66641]: value = "task-5145783" [ 706.525142] env[66641]: _type = "Task" [ 706.525142] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.537561] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145782, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.544858] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145783, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.585365] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Successfully created port: 5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 706.899418] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.899942] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.900444] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 706.900850] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 706.901115] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 706.904826] env[66641]: INFO nova.compute.manager [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Terminating instance [ 706.985685] env[66641]: DEBUG nova.scheduler.client.report [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 707.042174] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 707.044654] env[66641]: DEBUG oslo_vmware.api [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Task: {'id': task-5145782, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.29842} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.045606] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.045858] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.046118] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.046351] env[66641]: INFO nova.compute.manager [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Took 1.43 seconds to destroy the instance on the hypervisor. [ 707.046818] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 707.046903] env[66641]: DEBUG nova.compute.manager [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 707.047018] env[66641]: DEBUG nova.network.neutron [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 707.048466] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.048795] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.060684] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145783, 'name': PowerOffVM_Task, 'duration_secs': 0.31776} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.060980] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 707.061187] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 707.061472] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f4e69c94-10a8-4d12-84f6-84fd9b9104bc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.085745] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 707.087167] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 707.087167] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 707.087804] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 707.088062] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 707.088303] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 707.089102] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.089102] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 707.089102] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 707.089284] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 707.089519] env[66641]: DEBUG nova.virt.hardware [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 707.091197] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01a82c2-8a17-4f33-bd79-9c28548b6510 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.106277] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58fa9214-e1a9-4cc6-b852-1446389b63ff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.142476] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 707.142719] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 707.143042] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleting the datastore file [datastore2] d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 707.143389] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75aca930-1d68-4ddd-a9cb-6a764f3ea3e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.151420] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 707.151420] env[66641]: value = "task-5145785" [ 707.151420] env[66641]: _type = "Task" [ 707.151420] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.161981] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145785, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.411192] env[66641]: DEBUG nova.compute.manager [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 707.411192] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 707.412414] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f52440c4-2554-4066-99e7-d923f032439d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.423697] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 707.424520] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-594e0743-6345-49c7-82e1-d0e244237901 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.433926] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 707.433926] env[66641]: value = "task-5145786" [ 707.433926] env[66641]: _type = "Task" [ 707.433926] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.449938] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145786, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.492979] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.478s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 707.493581] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 707.496897] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.337s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 707.498937] env[66641]: INFO nova.compute.claims [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.667577] env[66641]: DEBUG oslo_vmware.api [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145785, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279764} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.668060] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 707.668964] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 707.668964] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 707.669124] env[66641]: INFO nova.compute.manager [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Took 1.18 seconds to destroy the instance on the hypervisor. [ 707.669470] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 707.669741] env[66641]: DEBUG nova.compute.manager [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 707.669935] env[66641]: DEBUG nova.network.neutron [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 707.670955] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.671221] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.879993] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 707.880441] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 707.945202] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145786, 'name': PowerOffVM_Task, 'duration_secs': 0.373607} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.945656] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 707.945730] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 707.946082] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da41dd55-328f-439f-9021-0036e88d6b3a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.004960] env[66641]: DEBUG nova.compute.utils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 708.011330] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 708.011330] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 708.011330] env[66641]: WARNING neutronclient.v2_0.client [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.011330] env[66641]: WARNING neutronclient.v2_0.client [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 708.011618] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.011876] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.032584] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 708.032896] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 708.033137] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleting the datastore file [datastore1] fe143d94-57ce-46d4-85e3-80d84af66dfc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 708.033541] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-637bd28e-368e-45d5-af04-2e3fc2a3c2bb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.039158] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.039444] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.049683] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 708.049683] env[66641]: value = "task-5145788" [ 708.049683] env[66641]: _type = "Task" [ 708.049683] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.064230] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145788, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.177721] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 708.178695] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 708.372382] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Successfully updated port: 5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 708.418376] env[66641]: DEBUG nova.policy [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '487c47af72b04d7fa50840a75de445ab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c44198647b5c4cc692c9972a7bf347c1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 708.511825] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 708.563421] env[66641]: DEBUG oslo_vmware.api [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145788, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208289} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.566745] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 708.567022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 708.567272] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 708.567482] env[66641]: INFO nova.compute.manager [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Took 1.16 seconds to destroy the instance on the hypervisor. [ 708.567755] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 708.568269] env[66641]: DEBUG nova.compute.manager [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 708.568409] env[66641]: DEBUG nova.network.neutron [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 708.569036] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.569343] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.681336] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 708.781258] env[66641]: DEBUG nova.network.neutron [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 708.796392] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 708.796392] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 708.825754] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Successfully created port: 43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 708.882720] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.882892] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 708.883076] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 708.925317] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df93d81-6ad4-44d9-9c58-302292405f3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.935147] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee2dfb20-ce15-47db-b13f-4aca7a326efa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.974814] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0a23ec8-8a42-4e6c-acb7-1517bbc927d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.985500] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b567feef-14f9-4166-993d-b475d25dd22f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.004275] env[66641]: DEBUG nova.compute.provider_tree [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.157892] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "a2d567c9-09af-44e1-aad2-c8f394a09714" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.157892] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.158117] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.158150] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 709.158330] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 709.164762] env[66641]: INFO nova.compute.manager [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Terminating instance [ 709.210201] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.266451] env[66641]: DEBUG nova.network.neutron [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 709.284191] env[66641]: INFO nova.compute.manager [-] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Took 2.24 seconds to deallocate network for instance. [ 709.387350] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.389848] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 709.394796] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 709.507942] env[66641]: DEBUG nova.scheduler.client.report [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 709.522469] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 709.579878] env[66641]: DEBUG nova.network.neutron [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 709.671825] env[66641]: DEBUG nova.compute.manager [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 709.671825] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.673290] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddd7e80b-2a69-4b67-bf7a-b6e072331398 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.683820] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 709.684039] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6a69ee7-dc07-476a-887e-4e8bc7714254 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.693403] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 709.693403] env[66641]: value = "task-5145789" [ 709.693403] env[66641]: _type = "Task" [ 709.693403] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.703884] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.730220] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 709.730533] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 709.730764] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 709.731099] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 709.731155] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 709.731337] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 709.731580] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.731800] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 709.732031] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 709.732238] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 709.732574] env[66641]: DEBUG nova.virt.hardware [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 709.733827] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b5ab9c-8f5d-4f7d-b50f-6d616d44634d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.743756] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd69182-236c-4e9d-b097-48fd3978e0b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.770324] env[66641]: INFO nova.compute.manager [-] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Took 2.10 seconds to deallocate network for instance. [ 709.792845] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 709.864135] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 709.864677] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 710.014133] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.517s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 710.014959] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 710.019709] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.924s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 710.019904] env[66641]: DEBUG nova.objects.instance [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lazy-loading 'resources' on Instance uuid 434e66b1-5ea7-43d2-9b41-bdeda2a312d7 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 710.083675] env[66641]: INFO nova.compute.manager [-] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Took 1.52 seconds to deallocate network for instance. [ 710.205646] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.280777] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.523566] env[66641]: DEBUG nova.compute.utils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 710.528381] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 710.592788] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 710.707109] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145789, 'name': PowerOffVM_Task, 'duration_secs': 0.698948} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.707376] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 710.707536] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 710.707798] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de8f56f0-17fd-4685-8fb8-76eafba1ba63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.760612] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Successfully updated port: 43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 710.782450] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 710.782717] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 710.784030] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Deleting the datastore file [datastore1] a2d567c9-09af-44e1-aad2-c8f394a09714 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 710.788568] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-49bc7288-ac4e-43ff-97d2-a0da09135adb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.802690] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for the task: (returnval){ [ 710.802690] env[66641]: value = "task-5145791" [ 710.802690] env[66641]: _type = "Task" [ 710.802690] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.817485] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.839434] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0f23e8-6d12-4e16-a990-b631a5c801e9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.851080] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef2ecd16-6c08-4d61-bcef-200c661c8569 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.889098] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e90fefa-55c4-420c-8a65-adce4cdc75ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.897844] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d6c2894-40c0-48ee-af4f-b8d1279178e8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.913188] env[66641]: DEBUG nova.compute.provider_tree [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.034030] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 711.141072] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.141752] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.264811] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.264901] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquired lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 711.267138] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 711.317406] env[66641]: DEBUG oslo_vmware.api [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Task: {'id': task-5145791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.172964} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.317868] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 711.318151] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 711.318424] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.318715] env[66641]: INFO nova.compute.manager [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Took 1.65 seconds to destroy the instance on the hypervisor. [ 711.319092] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 711.319362] env[66641]: DEBUG nova.compute.manager [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 711.319583] env[66641]: DEBUG nova.network.neutron [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 711.320166] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.320648] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.419485] env[66641]: DEBUG nova.scheduler.client.report [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 711.507166] env[66641]: DEBUG nova.network.neutron [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Updating instance_info_cache with network_info: [{"id": "5a4d98a2-57b2-40e0-9616-a27818d66651", "address": "fa:16:3e:de:40:68", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a4d98a2-57", "ovs_interfaceid": "5a4d98a2-57b2-40e0-9616-a27818d66651", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 711.541754] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.543059] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.772257] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 711.772634] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 711.777977] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 711.925452] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.906s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 711.929065] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.096s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 711.929466] env[66641]: DEBUG nova.objects.instance [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lazy-loading 'resources' on Instance uuid 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.966785] env[66641]: INFO nova.scheduler.client.report [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted allocations for instance 434e66b1-5ea7-43d2-9b41-bdeda2a312d7 [ 712.010474] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 712.011035] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Instance network_info: |[{"id": "5a4d98a2-57b2-40e0-9616-a27818d66651", "address": "fa:16:3e:de:40:68", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a4d98a2-57", "ovs_interfaceid": "5a4d98a2-57b2-40e0-9616-a27818d66651", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 712.011434] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:40:68', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5a4d98a2-57b2-40e0-9616-a27818d66651', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.023615] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Creating folder: Project (b2c8481015524aee95a933f61082faec). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.024748] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f01fb23c-fa09-47d9-ab58-60d7fa48719a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.040665] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Created folder: Project (b2c8481015524aee95a933f61082faec) in parent group-v1000566. [ 712.041127] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Creating folder: Instances. Parent ref: group-v1000635. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.041806] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-34d36796-38cb-4153-a485-fc47e3ed934c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.056896] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Created folder: Instances in parent group-v1000635. [ 712.057200] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 712.057420] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 712.057741] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b7395c2-1ecf-4f2d-bb8f-d4f0a6d41548 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.076028] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 712.085399] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 712.085399] env[66641]: value = "task-5145794" [ 712.085399] env[66641]: _type = "Task" [ 712.085399] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.095465] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145794, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.110240] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 712.111467] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.112210] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 712.112210] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 712.112210] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 712.112210] env[66641]: DEBUG nova.virt.hardware [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 712.117029] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb7f04f-2f96-4229-a8d5-b54589f35d24 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.123107] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b454c95-b89b-4a33-9a16-89c6a7fc15a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.141382] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 712.147413] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Creating folder: Project (0ad4b8d0c8b54e7c962aba80b138773b). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.147960] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-561f13d6-9874-4abf-a3cb-c74d7ebe97ba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.162412] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Created folder: Project (0ad4b8d0c8b54e7c962aba80b138773b) in parent group-v1000566. [ 712.163428] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Creating folder: Instances. Parent ref: group-v1000638. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 712.163428] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee7fc48c-67ae-4106-abf3-639bfc4af9d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.181059] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Created folder: Instances in parent group-v1000638. [ 712.181059] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 712.181437] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 712.181823] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-64d02695-a22a-4e15-b75c-f713176ca41b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.210138] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 712.210138] env[66641]: value = "task-5145797" [ 712.210138] env[66641]: _type = "Task" [ 712.210138] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.222296] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145797, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.366181] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.366730] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.428316] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 712.428991] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 712.484752] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4f87e2de-5c4a-4d1d-ae5c-98f18e799215 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "434e66b1-5ea7-43d2-9b41-bdeda2a312d7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.199s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.597830] env[66641]: DEBUG nova.network.neutron [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Updating instance_info_cache with network_info: [{"id": "43a23cc1-3c20-4707-9152-ad24a94338b5", "address": "fa:16:3e:09:2a:56", "network": {"id": "f4333c8e-9b5a-4509-b80a-f69274d0cec4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1631952192-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c44198647b5c4cc692c9972a7bf347c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "54c45719-5690-47bf-b45b-6cad9813071e", "external-id": "nsx-vlan-transportzone-62", "segmentation_id": 62, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a23cc1-3c", "ovs_interfaceid": "43a23cc1-3c20-4707-9152-ad24a94338b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 712.603320] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145794, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.726954] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145797, 'name': CreateVM_Task, 'duration_secs': 0.461918} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.731243] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 712.735387] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.735387] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 712.735387] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 712.735387] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2baf2b5-a7bd-4d63-95de-b63e5194fc7e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.740588] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 712.740588] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5222c343-4417-644b-2000-8554e4c30575" [ 712.740588] env[66641]: _type = "Task" [ 712.740588] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.755823] env[66641]: DEBUG nova.compute.manager [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Received event network-vif-plugged-5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 712.756147] env[66641]: DEBUG oslo_concurrency.lockutils [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] Acquiring lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 712.756435] env[66641]: DEBUG oslo_concurrency.lockutils [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 712.756690] env[66641]: DEBUG oslo_concurrency.lockutils [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 712.756931] env[66641]: DEBUG nova.compute.manager [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] No waiting events found dispatching network-vif-plugged-5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 712.757184] env[66641]: WARNING nova.compute.manager [req-b6a3bf31-51a6-4e9e-af05-22acf87543f1 req-79b50531-cb94-4e78-9181-c5443609d217 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Received unexpected event network-vif-plugged-5a4d98a2-57b2-40e0-9616-a27818d66651 for instance with vm_state building and task_state spawning. [ 712.766412] env[66641]: DEBUG nova.network.neutron [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 712.766760] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5222c343-4417-644b-2000-8554e4c30575, 'name': SearchDatastore_Task, 'duration_secs': 0.015831} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.771749] env[66641]: DEBUG nova.compute.manager [req-6fa1f634-1f0e-4574-8bbb-2b54f45ba5d9 req-91c23ba8-cac1-4d94-9240-1936748f98b2 service nova] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Received event network-vif-deleted-f4cb47fb-c2ae-459d-b011-8d21bfbc5a0a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 712.772650] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 712.773330] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.773548] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.775052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 712.775052] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.775052] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b14569a0-0184-470f-a4db-4f2e5d466fb4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.787796] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.788072] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 712.788872] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bdaaf4b6-5ecb-49fe-99d2-94e8f7917761 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.802203] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 712.802203] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523f1d20-00da-de8d-75c9-a858e6e8943e" [ 712.802203] env[66641]: _type = "Task" [ 712.802203] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.812920] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523f1d20-00da-de8d-75c9-a858e6e8943e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.816599] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8f4b147-b598-4587-b9e2-cef61b6fe340 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.825990] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ec32ed-0b9e-42c4-8777-ae8d44cc2b99 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.863873] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20560985-50bf-4d7c-aaa1-88b02259d7d2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.873681] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0d8fa3-9a32-497d-bd7c-aa100cc5a002 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.891397] env[66641]: DEBUG nova.compute.provider_tree [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.101523] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145794, 'name': CreateVM_Task, 'duration_secs': 0.537232} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.104022] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 713.104022] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.104022] env[66641]: WARNING openstack [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.113547] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.113547] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.113547] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 713.113547] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Releasing lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.113944] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Instance network_info: |[{"id": "43a23cc1-3c20-4707-9152-ad24a94338b5", "address": "fa:16:3e:09:2a:56", "network": {"id": "f4333c8e-9b5a-4509-b80a-f69274d0cec4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1631952192-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c44198647b5c4cc692c9972a7bf347c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "54c45719-5690-47bf-b45b-6cad9813071e", "external-id": "nsx-vlan-transportzone-62", "segmentation_id": 62, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a23cc1-3c", "ovs_interfaceid": "43a23cc1-3c20-4707-9152-ad24a94338b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 713.114276] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af84dda6-c082-432f-8369-ea2456054144 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.116413] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:09:2a:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '54c45719-5690-47bf-b45b-6cad9813071e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '43a23cc1-3c20-4707-9152-ad24a94338b5', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.124577] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Creating folder: Project (c44198647b5c4cc692c9972a7bf347c1). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.125488] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4ede5db-98ac-4910-9935-040f4f81fea4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.131642] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 713.131642] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5264db92-78a1-2df6-d61a-8ca6a19d4c05" [ 713.131642] env[66641]: _type = "Task" [ 713.131642] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.139985] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Created folder: Project (c44198647b5c4cc692c9972a7bf347c1) in parent group-v1000566. [ 713.140256] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Creating folder: Instances. Parent ref: group-v1000641. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 713.144619] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9bf33640-d198-4bd4-9563-fdabd9300710 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.146635] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5264db92-78a1-2df6-d61a-8ca6a19d4c05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.158707] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Created folder: Instances in parent group-v1000641. [ 713.158707] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 713.163040] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 713.167116] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f91ce2c8-1dca-4b16-9873-1c8c32a54c42 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.189432] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "b9d032da-031e-42e0-86e2-95254c1ceac1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.189432] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.197673] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.197673] env[66641]: value = "task-5145800" [ 713.197673] env[66641]: _type = "Task" [ 713.197673] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.208802] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145800, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.267771] env[66641]: INFO nova.compute.manager [-] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Took 1.95 seconds to deallocate network for instance. [ 713.315609] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523f1d20-00da-de8d-75c9-a858e6e8943e, 'name': SearchDatastore_Task, 'duration_secs': 0.013498} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.317485] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8225901b-8155-4034-9020-e7954bb6cdaf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.325773] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 713.325773] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522afaf8-4cb6-cbd8-9908-c132bc86ac3b" [ 713.325773] env[66641]: _type = "Task" [ 713.325773] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.336664] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522afaf8-4cb6-cbd8-9908-c132bc86ac3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.395395] env[66641]: DEBUG nova.scheduler.client.report [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 713.473043] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 713.474436] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c04a142-c267-4453-930b-543d40c5d8b5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.481414] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 713.481576] env[66641]: ERROR oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk due to incomplete transfer. [ 713.481872] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-dfee1207-d0af-4c7a-819b-57c093291ec4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.496504] env[66641]: DEBUG oslo_vmware.rw_handles [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52845d55-7a6b-4324-4f70-52c148adadb0/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 713.497175] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Uploaded image b6f3d72e-8bf4-4ee3-a9f9-3195a8e0534e to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 713.500288] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 713.501162] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ea3232f6-24bc-411d-b5ba-7d4fa0ff90a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.512138] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 713.512138] env[66641]: value = "task-5145801" [ 713.512138] env[66641]: _type = "Task" [ 713.512138] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.526157] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145801, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.643402] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5264db92-78a1-2df6-d61a-8ca6a19d4c05, 'name': SearchDatastore_Task, 'duration_secs': 0.016025} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.643731] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.644019] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 713.644775] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.694106] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 713.711633] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145800, 'name': CreateVM_Task, 'duration_secs': 0.385255} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.712033] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 713.713223] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 713.714308] env[66641]: WARNING openstack [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 713.722437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.722437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.722437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 713.722950] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b3d3824a-cab8-4c7e-ab85-02e72daa6c7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.729937] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 713.729937] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e08df5-3602-7a18-1a1a-04e0026fa47d" [ 713.729937] env[66641]: _type = "Task" [ 713.729937] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.744230] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e08df5-3602-7a18-1a1a-04e0026fa47d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.779243] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 713.838647] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522afaf8-4cb6-cbd8-9908-c132bc86ac3b, 'name': SearchDatastore_Task, 'duration_secs': 0.011491} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.838647] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 713.838647] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 713.838647] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 713.839119] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 713.839119] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5b0084d-b1d3-47f3-9ec2-ac05393595df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.840766] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-df2b8a9f-894f-482a-8d11-378a3d665ca3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.848893] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 713.848893] env[66641]: value = "task-5145802" [ 713.848893] env[66641]: _type = "Task" [ 713.848893] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.853449] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 713.853621] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 713.854769] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d3c2e0b-f248-4275-825b-c6ef8805a91e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.860385] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145802, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.863989] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 713.863989] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524bd910-7c14-9e0e-295f-e8953a168103" [ 713.863989] env[66641]: _type = "Task" [ 713.863989] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.872365] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524bd910-7c14-9e0e-295f-e8953a168103, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.900595] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.972s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 713.904117] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.693s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 713.904874] env[66641]: INFO nova.compute.claims [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.944186] env[66641]: INFO nova.scheduler.client.report [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Deleted allocations for instance 5963a8ba-9d2c-42c3-a5da-25c29bf9e763 [ 714.025496] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145801, 'name': Destroy_Task, 'duration_secs': 0.366568} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.025772] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Destroyed the VM [ 714.026011] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 714.026286] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9ba01d2f-3bf5-442f-8e72-27bd9d6034c2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.035393] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 714.035393] env[66641]: value = "task-5145803" [ 714.035393] env[66641]: _type = "Task" [ 714.035393] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.045856] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145803, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.220423] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.250523] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e08df5-3602-7a18-1a1a-04e0026fa47d, 'name': SearchDatastore_Task, 'duration_secs': 0.034207} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.251130] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.251540] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.252079] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.363251] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145802, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.379481] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524bd910-7c14-9e0e-295f-e8953a168103, 'name': SearchDatastore_Task, 'duration_secs': 0.040267} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.380355] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b87d4e6c-c722-43b0-bf72-41f83bc1bded {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.388785] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 714.388785] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5261f3e9-90c0-5eb6-f523-f766270b606a" [ 714.388785] env[66641]: _type = "Task" [ 714.388785] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.405845] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5261f3e9-90c0-5eb6-f523-f766270b606a, 'name': SearchDatastore_Task, 'duration_secs': 0.010922} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.406179] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 714.406529] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 3619be34-19ed-4d3d-b3ee-573126dcefbb/3619be34-19ed-4d3d-b3ee-573126dcefbb.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 714.406850] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 714.407255] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.407314] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2b52cb8-b8e1-487f-9789-493f2c9ac49c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.409860] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f064dcbf-6738-4ace-96d3-d013411f2c0a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.421200] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 714.421200] env[66641]: value = "task-5145804" [ 714.421200] env[66641]: _type = "Task" [ 714.421200] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.432857] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.435387] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 714.435387] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 714.436595] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1a5b3db-f7d9-4b16-aa65-2c09d57af623 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.445079] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 714.445079] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d3d2f9-7c1d-7ae4-4a16-0e45648bab98" [ 714.445079] env[66641]: _type = "Task" [ 714.445079] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.460878] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d3d2f9-7c1d-7ae4-4a16-0e45648bab98, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.460878] env[66641]: DEBUG oslo_concurrency.lockutils [None req-274d1de6-e3d5-4a07-aeec-7ee2ce19b060 tempest-ServersAdmin275Test-1727789183 tempest-ServersAdmin275Test-1727789183-project-member] Lock "5963a8ba-9d2c-42c3-a5da-25c29bf9e763" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.570s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.539552] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.539850] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.551377] env[66641]: DEBUG oslo_vmware.api [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145803, 'name': RemoveSnapshot_Task, 'duration_secs': 0.465954} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.551665] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 714.553099] env[66641]: INFO nova.compute.manager [None req-d0b21102-0107-4d92-a1c7-8591c246aa04 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Took 17.08 seconds to snapshot the instance on the hypervisor. [ 714.599524] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.600023] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.002s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.600533] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 714.600533] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 714.600813] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 714.603871] env[66641]: INFO nova.compute.manager [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Terminating instance [ 714.863889] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145802, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5614} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.863889] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 714.863889] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 714.863889] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c37a0be-9fea-4049-8d0d-55eff3edcbde {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.871022] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 714.871022] env[66641]: value = "task-5145805" [ 714.871022] env[66641]: _type = "Task" [ 714.871022] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.879600] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145805, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.934853] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.961026] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d3d2f9-7c1d-7ae4-4a16-0e45648bab98, 'name': SearchDatastore_Task, 'duration_secs': 0.011243} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.964802] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b670402a-5413-4bf9-9d09-21441d79e6b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.971779] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 714.971779] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52904d72-e55b-1109-a657-544e4cd743fb" [ 714.971779] env[66641]: _type = "Task" [ 714.971779] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.983961] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52904d72-e55b-1109-a657-544e4cd743fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.045026] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 715.108360] env[66641]: DEBUG nova.compute.manager [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 715.108592] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 715.112096] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2345819a-05d8-4e5d-bcf9-bebecae9acac {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.121089] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 715.121419] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9fd3983a-75ca-4395-929c-04df14ad92ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.133478] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 715.133478] env[66641]: value = "task-5145806" [ 715.133478] env[66641]: _type = "Task" [ 715.133478] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.143976] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145806, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.208964] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9257dc7-d92b-4c9a-af06-fa1120f8236c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.221016] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e770c8da-5fab-4513-b155-03859666ee82 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.258465] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5c6b6b-854b-4e84-9cc6-64bbff4cc48d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.266845] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34e27fe9-a204-47c7-adea-74dd7e276003 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.281693] env[66641]: DEBUG nova.compute.provider_tree [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.379628] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145805, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075789} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.380168] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 715.381054] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fde0b01-f497-4305-94b6-63dee34e5ec7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.403182] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Reconfiguring VM instance instance-00000019 to attach disk [datastore2] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 715.403505] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b06cd3a-349b-4b71-821b-1190e6603893 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.424777] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 715.424777] env[66641]: value = "task-5145807" [ 715.424777] env[66641]: _type = "Task" [ 715.424777] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.440907] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145807, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.441110] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145804, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.489287] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52904d72-e55b-1109-a657-544e4cd743fb, 'name': SearchDatastore_Task, 'duration_secs': 0.010352} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.489601] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 715.489901] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 8235441c-c046-49da-a5a5-92bd85e17982/8235441c-c046-49da-a5a5-92bd85e17982.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 715.490198] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce001d22-f4fe-46da-8912-1e10dcfeed66 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.499146] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 715.499146] env[66641]: value = "task-5145808" [ 715.499146] env[66641]: _type = "Task" [ 715.499146] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.509717] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.574569] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 715.645301] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145806, 'name': PowerOffVM_Task, 'duration_secs': 0.230084} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.645938] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 715.646218] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 715.647558] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-da0bc1a0-20c7-48cf-9219-b378f07bcba6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.725072] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 715.725374] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 715.725552] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Deleting the datastore file [datastore1] ac5bb5a2-e630-42f0-be0c-24f52be81367 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 715.725880] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50baae63-5e12-44d5-b3d8-3bc9a71f794f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.735286] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for the task: (returnval){ [ 715.735286] env[66641]: value = "task-5145810" [ 715.735286] env[66641]: _type = "Task" [ 715.735286] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.747155] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.785886] env[66641]: DEBUG nova.scheduler.client.report [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 715.946932] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145804, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.50166} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.948665] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 3619be34-19ed-4d3d-b3ee-573126dcefbb/3619be34-19ed-4d3d-b3ee-573126dcefbb.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 715.948888] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 715.950121] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145807, 'name': ReconfigVM_Task, 'duration_secs': 0.344525} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.950349] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-46ddd802-580c-4852-91ac-b3f6ab618a7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.953872] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Reconfigured VM instance instance-00000019 to attach disk [datastore2] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.954545] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a531ea21-1bdf-4599-8fe1-e52ff3aa95e2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.962879] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 715.962879] env[66641]: value = "task-5145812" [ 715.962879] env[66641]: _type = "Task" [ 715.962879] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.964723] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 715.964723] env[66641]: value = "task-5145811" [ 715.964723] env[66641]: _type = "Task" [ 715.964723] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.980561] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145812, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.984870] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145811, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.011201] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145808, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.259841] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.293572] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 716.294326] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 716.304234] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.509s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 716.304234] env[66641]: DEBUG nova.objects.instance [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lazy-loading 'resources' on Instance uuid 22899269-28fc-446e-9533-1f6862a9e8db {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 716.484079] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145812, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082809} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.491502] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 716.492439] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145811, 'name': Rename_Task, 'duration_secs': 0.264027} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.493886] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10fb0e9-944f-4c59-9389-bfe2dc0ea4eb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.497160] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 716.498122] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acef340d-480b-4724-ad3b-15b269ba605d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.526764] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Reconfiguring VM instance instance-00000017 to attach disk [datastore2] 3619be34-19ed-4d3d-b3ee-573126dcefbb/3619be34-19ed-4d3d-b3ee-573126dcefbb.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 716.529230] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea68cf6d-731f-4369-9c04-00b682454032 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.545732] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 716.545732] env[66641]: value = "task-5145813" [ 716.545732] env[66641]: _type = "Task" [ 716.545732] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.554220] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145808, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.980328} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.555096] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 8235441c-c046-49da-a5a5-92bd85e17982/8235441c-c046-49da-a5a5-92bd85e17982.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 716.555096] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.556967] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9f44c216-6f11-4e65-83e4-8a84672da855 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.559327] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 716.559327] env[66641]: value = "task-5145814" [ 716.559327] env[66641]: _type = "Task" [ 716.559327] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.568170] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145813, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.573181] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 716.573181] env[66641]: value = "task-5145815" [ 716.573181] env[66641]: _type = "Task" [ 716.573181] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.573611] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145814, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.583863] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145815, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.748724] env[66641]: DEBUG oslo_vmware.api [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Task: {'id': task-5145810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.542832} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.749238] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 716.749553] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 716.750720] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 716.750720] env[66641]: INFO nova.compute.manager [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Took 1.64 seconds to destroy the instance on the hypervisor. [ 716.750720] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 716.750720] env[66641]: DEBUG nova.compute.manager [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 716.750720] env[66641]: DEBUG nova.network.neutron [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 716.751421] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.751511] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.809020] env[66641]: DEBUG nova.compute.utils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 716.815987] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 716.815987] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 716.815987] env[66641]: WARNING neutronclient.v2_0.client [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.815987] env[66641]: WARNING neutronclient.v2_0.client [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 716.815987] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.816215] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.880488] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 716.880777] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 716.987442] env[66641]: DEBUG nova.policy [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7d6c5fbc65414faebe5c3c9bb4ec4309', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ab82c335b15478788a818319bf11c03', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 717.062522] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145813, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.075081] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145814, 'name': ReconfigVM_Task, 'duration_secs': 0.439007} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.079905] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Reconfigured VM instance instance-00000017 to attach disk [datastore2] 3619be34-19ed-4d3d-b3ee-573126dcefbb/3619be34-19ed-4d3d-b3ee-573126dcefbb.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 717.083770] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10c5c8d8-9308-4ba6-bf41-07a4c23434af {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.096198] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145815, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.227651} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.096522] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 717.096522] env[66641]: value = "task-5145816" [ 717.096522] env[66641]: _type = "Task" [ 717.096522] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.096798] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.097775] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b73b414b-e880-476b-8560-2eec422e05dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.112591] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145816, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.133120] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Reconfiguring VM instance instance-00000018 to attach disk [datastore2] 8235441c-c046-49da-a5a5-92bd85e17982/8235441c-c046-49da-a5a5-92bd85e17982.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.136159] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4d09be6d-a7c6-4774-87fc-9ad227aebbf1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.166989] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 717.166989] env[66641]: value = "task-5145817" [ 717.166989] env[66641]: _type = "Task" [ 717.166989] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.175026] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f602717f-f72b-420a-8ad0-df4c34439b4a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.184135] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.188478] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6bcdb94-1047-4718-bf59-de5ce7d4e2bf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.228200] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-081c0afc-3599-4ce2-b7db-efe2d8bf60c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.238402] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8a7266-2eca-41e1-a6ed-a9b6e82221e4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.256346] env[66641]: DEBUG nova.compute.provider_tree [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 717.315276] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 717.452634] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Successfully created port: 6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 717.569518] env[66641]: DEBUG oslo_vmware.api [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145813, 'name': PowerOnVM_Task, 'duration_secs': 0.623665} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.569518] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 717.569846] env[66641]: INFO nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Took 5.49 seconds to spawn the instance on the hypervisor. [ 717.569940] env[66641]: DEBUG nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 717.570843] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa856376-2535-423e-852a-de44a6ca7f5a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.606582] env[66641]: DEBUG nova.compute.manager [req-1b79d3cb-a1de-4691-b863-56d082b48747 req-92880313-5699-4336-b8d5-cf6245384382 service nova] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Received event network-vif-deleted-aa8a5251-c99e-43d0-8dfc-6c4476b6a0e4 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 717.614884] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145816, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.651204] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Received event network-changed-5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 717.651204] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Refreshing instance network info cache due to event network-changed-5a4d98a2-57b2-40e0-9616-a27818d66651. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 717.651375] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Acquiring lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.651630] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Acquired lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 717.651630] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Refreshing network info cache for port 5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 717.682623] env[66641]: DEBUG nova.network.neutron [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 717.687131] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145817, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.760201] env[66641]: DEBUG nova.scheduler.client.report [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 718.040221] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.040571] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.040879] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.041172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.041360] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.044218] env[66641]: INFO nova.compute.manager [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Terminating instance [ 718.098890] env[66641]: INFO nova.compute.manager [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Took 16.97 seconds to build instance. [ 718.112349] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145816, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.155112] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 718.155451] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 718.175789] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145817, 'name': ReconfigVM_Task, 'duration_secs': 0.650918} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.176083] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Reconfigured VM instance instance-00000018 to attach disk [datastore2] 8235441c-c046-49da-a5a5-92bd85e17982/8235441c-c046-49da-a5a5-92bd85e17982.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.176781] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef782f88-f78c-41db-bf1a-8ce4b670fef8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.184251] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 718.184251] env[66641]: value = "task-5145818" [ 718.184251] env[66641]: _type = "Task" [ 718.184251] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.193548] env[66641]: INFO nova.compute.manager [-] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Took 1.44 seconds to deallocate network for instance. [ 718.193874] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145818, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.267997] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.269984] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.989s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 718.270241] env[66641]: DEBUG nova.objects.instance [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lazy-loading 'resources' on Instance uuid d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 718.319614] env[66641]: INFO nova.scheduler.client.report [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Deleted allocations for instance 22899269-28fc-446e-9533-1f6862a9e8db [ 718.326322] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 718.355717] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 718.355963] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 718.356133] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 718.356311] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 718.356448] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 718.356587] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 718.356789] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 718.356941] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 718.357123] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 718.357282] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 718.357446] env[66641]: DEBUG nova.virt.hardware [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 718.358381] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755f3238-865a-49be-bb8e-4e2fff9c58a9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.367603] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ba57c5-b917-43c9-ae8b-6e700c854d24 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.548171] env[66641]: DEBUG nova.compute.manager [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 718.548427] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.549547] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45a2d8d-5aeb-4a8a-999b-05d5d90c2e4a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.560874] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 718.561320] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf59dc4b-e972-4950-9b06-c270cc1b8289 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.570593] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 718.570593] env[66641]: value = "task-5145819" [ 718.570593] env[66641]: _type = "Task" [ 718.570593] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.584671] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.603798] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cfb83fad-3031-403d-bc5c-02ef53748634 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.495s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 718.616573] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145816, 'name': Rename_Task, 'duration_secs': 1.181318} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.617165] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 718.617486] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0091379-ac2c-40a3-9656-fc0430fe75fc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.625835] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 718.625835] env[66641]: value = "task-5145820" [ 718.625835] env[66641]: _type = "Task" [ 718.625835] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.636133] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145820, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.697884] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145818, 'name': Rename_Task, 'duration_secs': 0.327069} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.697884] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 718.697884] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a2a4667a-0fc6-49a6-bb4e-6101ee7a7e52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.701336] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 718.705036] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 718.705036] env[66641]: value = "task-5145821" [ 718.705036] env[66641]: _type = "Task" [ 718.705036] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.714297] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.833501] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0678ef73-0270-4ab0-af0d-f9edcf75ac23 tempest-DeleteServersAdminTestJSON-222019155 tempest-DeleteServersAdminTestJSON-222019155-project-member] Lock "22899269-28fc-446e-9533-1f6862a9e8db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.727s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 719.027320] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6ff5e21-4e9d-4b0b-af26-e64374d28364 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.049195] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f7301d-abef-43e4-86e3-b8f7023a20a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.092440] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d986c7ec-9504-4ac3-ab7f-68afd06a2338 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.103871] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145819, 'name': PowerOffVM_Task, 'duration_secs': 0.251869} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.106769] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 719.106874] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 719.107289] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-57e4610d-5cb6-4f37-a7bf-9f91dc81e5f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.110963] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1a5f552-3b3b-4127-987e-e2d37060030b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.132103] env[66641]: DEBUG nova.compute.provider_tree [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 719.145469] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145820, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.211837] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 719.212218] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 719.212512] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Deleting the datastore file [datastore1] eeb2556a-4e0c-43d7-83bd-942be5d5fdd2 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 719.219099] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87a57244-e2d0-474b-9dac-189db23da002 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.228625] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145821, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.230228] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for the task: (returnval){ [ 719.230228] env[66641]: value = "task-5145823" [ 719.230228] env[66641]: _type = "Task" [ 719.230228] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.241527] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145823, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.280977] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Successfully updated port: 6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 719.642676] env[66641]: DEBUG nova.scheduler.client.report [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 719.646261] env[66641]: DEBUG oslo_vmware.api [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145820, 'name': PowerOnVM_Task, 'duration_secs': 0.558107} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.646818] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.647030] env[66641]: INFO nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Took 12.60 seconds to spawn the instance on the hypervisor. [ 719.647206] env[66641]: DEBUG nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 719.648182] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338a8468-1095-4dfe-bbf1-ec8e855d90c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.719440] env[66641]: DEBUG oslo_vmware.api [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145821, 'name': PowerOnVM_Task, 'duration_secs': 0.780007} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.719954] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 719.720156] env[66641]: INFO nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Took 10.20 seconds to spawn the instance on the hypervisor. [ 719.720352] env[66641]: DEBUG nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 719.721294] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4096726-121b-4c37-bcb1-08ecd34cd466 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.746399] env[66641]: DEBUG oslo_vmware.api [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Task: {'id': task-5145823, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.284118} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.746686] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 719.746890] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 719.747106] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 719.747281] env[66641]: INFO nova.compute.manager [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Took 1.20 seconds to destroy the instance on the hypervisor. [ 719.747546] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 719.747751] env[66641]: DEBUG nova.compute.manager [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 719.747870] env[66641]: DEBUG nova.network.neutron [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 719.748742] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 719.749022] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 719.785759] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.786027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 719.786154] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 720.108875] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.109265] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.150182] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.878s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.152528] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.153206] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.160310] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.568s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 720.161098] env[66641]: DEBUG nova.objects.instance [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lazy-loading 'resources' on Instance uuid fe143d94-57ce-46d4-85e3-80d84af66dfc {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 720.176805] env[66641]: INFO nova.compute.manager [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Took 20.33 seconds to build instance. [ 720.203646] env[66641]: INFO nova.scheduler.client.report [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleted allocations for instance d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3 [ 720.243566] env[66641]: INFO nova.compute.manager [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Took 19.24 seconds to build instance. [ 720.294022] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 720.294022] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 720.298599] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 720.680965] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d4897985-e966-4343-932d-cbd3034d298d tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.863s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.711876] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a74b9f06-f742-4035-a151-4b7135c9e0ac tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.736s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.746540] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1cc9a33e-a658-4c15-a755-466188c8f368 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.777s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 720.881108] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94da9b11-4ccf-4688-939d-35997306fc89 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.893467] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58789ae5-7cbb-42da-be4d-4b2830c24c30 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.923224] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c97a459-02ec-4f05-bf4d-b5a9afb631b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.932085] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6afe76b-2a24-4c37-9647-d4b17ab3ab0c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.946819] env[66641]: DEBUG nova.compute.provider_tree [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 721.452704] env[66641]: DEBUG nova.scheduler.client.report [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 721.951319] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.951319] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.960076] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.800s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 721.962176] env[66641]: DEBUG nova.network.neutron [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 721.965179] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 721.965557] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 721.972817] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.194s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 721.973151] env[66641]: DEBUG nova.objects.instance [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lazy-loading 'resources' on Instance uuid a2d567c9-09af-44e1-aad2-c8f394a09714 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 721.997205] env[66641]: INFO nova.scheduler.client.report [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted allocations for instance fe143d94-57ce-46d4-85e3-80d84af66dfc [ 722.476537] env[66641]: INFO nova.compute.manager [-] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Took 2.73 seconds to deallocate network for instance. [ 722.508829] env[66641]: DEBUG oslo_concurrency.lockutils [None req-39f971b5-742d-4ee2-aa2a-b0d200de4528 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "fe143d94-57ce-46d4-85e3-80d84af66dfc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.609s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 722.778576] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3016eb6c-ffbe-4726-b4b3-359ebc3c2825 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.788492] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59f0e32a-3d21-444e-b432-693a307bde03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.819787] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96fe5ba7-d968-43cf-aa79-190378d5d6a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.828652] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b03a34-d0c9-4ece-b45a-a147ae174d38 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.843406] env[66641]: DEBUG nova.compute.provider_tree [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.986834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.349863] env[66641]: DEBUG nova.scheduler.client.report [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 723.496342] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Updated VIF entry in instance network info cache for port 5a4d98a2-57b2-40e0-9616-a27818d66651. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 723.496699] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Updating instance_info_cache with network_info: [{"id": "5a4d98a2-57b2-40e0-9616-a27818d66651", "address": "fa:16:3e:de:40:68", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5a4d98a2-57", "ovs_interfaceid": "5a4d98a2-57b2-40e0-9616-a27818d66651", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.531565] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 723.532014] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 723.748094] env[66641]: DEBUG nova.compute.manager [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Received event network-vif-plugged-6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 723.748321] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Acquiring lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 723.749238] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.749238] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.749238] env[66641]: DEBUG nova.compute.manager [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] No waiting events found dispatching network-vif-plugged-6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 723.749238] env[66641]: WARNING nova.compute.manager [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Received unexpected event network-vif-plugged-6d596bf0-768d-4562-9bc7-6349b3d94101 for instance with vm_state building and task_state spawning. [ 723.749238] env[66641]: DEBUG nova.compute.manager [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Received event network-changed-6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 723.749502] env[66641]: DEBUG nova.compute.manager [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Refreshing instance network info cache due to event network-changed-6d596bf0-768d-4562-9bc7-6349b3d94101. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 723.749502] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Acquiring lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.798495] env[66641]: DEBUG nova.network.neutron [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Updating instance_info_cache with network_info: [{"id": "6d596bf0-768d-4562-9bc7-6349b3d94101", "address": "fa:16:3e:3a:a5:88", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.122", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d596bf0-76", "ovs_interfaceid": "6d596bf0-768d-4562-9bc7-6349b3d94101", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 723.856858] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.883s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 723.859104] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.638s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 723.861661] env[66641]: INFO nova.compute.claims [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.897033] env[66641]: INFO nova.scheduler.client.report [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Deleted allocations for instance a2d567c9-09af-44e1-aad2-c8f394a09714 [ 724.004576] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Releasing lock "refresh_cache-3619be34-19ed-4d3d-b3ee-573126dcefbb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.004576] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Received event network-vif-deleted-71cb13b6-7303-4801-a446-6913f2523c32 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 724.004576] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Received event network-vif-plugged-43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 724.004576] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Acquiring lock "8235441c-c046-49da-a5a5-92bd85e17982-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 724.004576] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Lock "8235441c-c046-49da-a5a5-92bd85e17982-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 724.004958] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Lock "8235441c-c046-49da-a5a5-92bd85e17982-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.004958] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] No waiting events found dispatching network-vif-plugged-43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 724.004958] env[66641]: WARNING nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Received unexpected event network-vif-plugged-43a23cc1-3c20-4707-9152-ad24a94338b5 for instance with vm_state building and task_state spawning. [ 724.004958] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Received event network-changed-43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 724.004958] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Refreshing instance network info cache due to event network-changed-43a23cc1-3c20-4707-9152-ad24a94338b5. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 724.005120] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Acquiring lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.005120] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Acquired lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.005120] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Refreshing network info cache for port 43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 724.166791] env[66641]: DEBUG nova.compute.manager [req-9e99fa4a-36f0-447e-af04-2722a25edab4 req-247dcadf-3b94-417a-8557-0b4031ff9cd1 service nova] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Received event network-vif-deleted-33cb5bcf-2ebf-4fcc-8802-c4d6ad2be8f0 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 724.301523] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 724.301920] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Instance network_info: |[{"id": "6d596bf0-768d-4562-9bc7-6349b3d94101", "address": "fa:16:3e:3a:a5:88", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.122", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d596bf0-76", "ovs_interfaceid": "6d596bf0-768d-4562-9bc7-6349b3d94101", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 724.302989] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Acquired lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.302989] env[66641]: DEBUG nova.network.neutron [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Refreshing network info cache for port 6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 724.303699] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:a5:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6d596bf0-768d-4562-9bc7-6349b3d94101', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.312602] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 724.312940] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 724.313255] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65354f71-e08f-4a34-a3bb-238b0dec6fb6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.337028] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.337028] env[66641]: value = "task-5145824" [ 724.337028] env[66641]: _type = "Task" [ 724.337028] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.347108] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145824, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.409863] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c299499b-b069-4089-ac7f-013bf4d3f510 tempest-ImagesNegativeTestJSON-237239980 tempest-ImagesNegativeTestJSON-237239980-project-member] Lock "a2d567c9-09af-44e1-aad2-c8f394a09714" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.252s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 724.509984] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.510412] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.818772] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.818772] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.849013] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145824, 'name': CreateVM_Task, 'duration_secs': 0.409003} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.849211] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 724.850098] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 724.850997] env[66641]: WARNING openstack [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 724.856016] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.856164] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 724.856702] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 724.856806] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edf2378e-2610-4a3f-aa5a-060905c7fbcd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.862966] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 724.862966] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5221b15a-37d6-7fa5-db2e-21e24434c965" [ 724.862966] env[66641]: _type = "Task" [ 724.862966] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.873501] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5221b15a-37d6-7fa5-db2e-21e24434c965, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.108891] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68f4b540-c258-4cd0-b27a-25eb90c02945 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.119515] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62265b97-cef8-422a-898a-ee9968b3066d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.166856] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e64333-ac63-41ad-8fc3-457bee78550e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.176618] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6fe8c1-d51e-4e11-b0db-edf3f6f071f3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.191924] env[66641]: DEBUG nova.compute.provider_tree [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.377988] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5221b15a-37d6-7fa5-db2e-21e24434c965, 'name': SearchDatastore_Task, 'duration_secs': 0.018931} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.378338] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 725.378625] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 725.379213] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.379213] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 725.379213] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 725.379487] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4f16b468-3bf7-4852-b9ad-624d42b0d250 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.394568] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 725.394762] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 725.395566] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f24f9744-e278-48b6-bcc9-424d6c404f13 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.403147] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 725.403147] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52044929-fe1e-92b0-04af-ce93fc24243b" [ 725.403147] env[66641]: _type = "Task" [ 725.403147] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.414441] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52044929-fe1e-92b0-04af-ce93fc24243b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.579297] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 725.579297] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 725.695680] env[66641]: DEBUG nova.scheduler.client.report [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 725.915455] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52044929-fe1e-92b0-04af-ce93fc24243b, 'name': SearchDatastore_Task, 'duration_secs': 0.025898} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.915703] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-191ebf5b-1c09-4814-82fa-c7dcf0fc5565 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.921983] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 725.921983] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc8525-c0af-acc5-aba9-bf9a6af14be8" [ 725.921983] env[66641]: _type = "Task" [ 725.921983] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.930742] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc8525-c0af-acc5-aba9-bf9a6af14be8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.201560] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.343s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 726.202203] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 726.205053] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.631s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 726.206523] env[66641]: INFO nova.compute.claims [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 726.390931] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.391367] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.433125] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc8525-c0af-acc5-aba9-bf9a6af14be8, 'name': SearchDatastore_Task, 'duration_secs': 0.0435} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.434600] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.435106] env[66641]: WARNING openstack [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.442624] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 726.442897] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 530f1e0e-5911-434e-bf20-edfd9778d7ad/530f1e0e-5911-434e-bf20-edfd9778d7ad.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 726.443476] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e7b6c504-628a-4d78-9a2c-41cd352084f3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.451295] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 726.451295] env[66641]: value = "task-5145825" [ 726.451295] env[66641]: _type = "Task" [ 726.451295] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 726.462936] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145825, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.716081] env[66641]: DEBUG nova.compute.utils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 726.718650] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 726.718650] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 726.718650] env[66641]: WARNING neutronclient.v2_0.client [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.718650] env[66641]: WARNING neutronclient.v2_0.client [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 726.719584] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 726.719584] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 726.962420] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145825, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.161213] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.161543] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.161742] env[66641]: DEBUG nova.compute.manager [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 727.164382] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1b28c3a-6ccc-4a75-a003-0b8bf39376ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.171725] env[66641]: DEBUG nova.compute.manager [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=66641) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3459}} [ 727.172432] env[66641]: DEBUG nova.objects.instance [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lazy-loading 'flavor' on Instance uuid 3619be34-19ed-4d3d-b3ee-573126dcefbb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 727.182629] env[66641]: DEBUG nova.policy [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e6189ba08b374d808d7d565c8998b54f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04f038652b90481f9629847d20f93f84', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 727.226684] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 727.226684] env[66641]: WARNING openstack [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 727.235255] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 727.304902] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Updated VIF entry in instance network info cache for port 43a23cc1-3c20-4707-9152-ad24a94338b5. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 727.305285] env[66641]: DEBUG nova.network.neutron [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Updating instance_info_cache with network_info: [{"id": "43a23cc1-3c20-4707-9152-ad24a94338b5", "address": "fa:16:3e:09:2a:56", "network": {"id": "f4333c8e-9b5a-4509-b80a-f69274d0cec4", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-1631952192-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "c44198647b5c4cc692c9972a7bf347c1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "54c45719-5690-47bf-b45b-6cad9813071e", "external-id": "nsx-vlan-transportzone-62", "segmentation_id": 62, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43a23cc1-3c", "ovs_interfaceid": "43a23cc1-3c20-4707-9152-ad24a94338b5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.465529] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145825, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692219} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.468474] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 530f1e0e-5911-434e-bf20-edfd9778d7ad/530f1e0e-5911-434e-bf20-edfd9778d7ad.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 727.468688] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 727.469245] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f034fb34-f55a-4f60-ab97-56ba42ca77fe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.476902] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 727.476902] env[66641]: value = "task-5145826" [ 727.476902] env[66641]: _type = "Task" [ 727.476902] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 727.492323] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145826, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.508565] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dba16f98-4b53-4761-95bc-0af2a4d8c6a6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.516906] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe571b7-4e19-42e0-ba9b-6e3f0419ead3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.560538] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718172df-cc66-410a-b5f1-a98a09d41294 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.569143] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef9f31f2-6f51-472d-a837-96d01116a73c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.585476] env[66641]: DEBUG nova.compute.provider_tree [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.748463] env[66641]: DEBUG nova.network.neutron [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Updated VIF entry in instance network info cache for port 6d596bf0-768d-4562-9bc7-6349b3d94101. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 727.748822] env[66641]: DEBUG nova.network.neutron [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Updating instance_info_cache with network_info: [{"id": "6d596bf0-768d-4562-9bc7-6349b3d94101", "address": "fa:16:3e:3a:a5:88", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.122", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6d596bf0-76", "ovs_interfaceid": "6d596bf0-768d-4562-9bc7-6349b3d94101", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 727.769097] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Successfully created port: 0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 727.808330] env[66641]: DEBUG oslo_concurrency.lockutils [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] Releasing lock "refresh_cache-8235441c-c046-49da-a5a5-92bd85e17982" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 727.809913] env[66641]: DEBUG nova.compute.manager [req-f10d6712-d360-4cdc-886e-f297674cfd61 req-54837831-10bd-4431-b683-0afa85de91e7 service nova] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Received event network-vif-deleted-d66320d7-4884-46e8-83ab-a8a6fd0b09a7 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 727.858424] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.858696] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.858902] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 727.859707] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 727.859707] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 727.866126] env[66641]: INFO nova.compute.manager [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Terminating instance [ 727.987945] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145826, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082066} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 727.988374] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 727.989610] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34383b18-f6b8-4530-8b10-93f1fb98d716 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.022397] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Reconfiguring VM instance instance-0000001a to attach disk [datastore2] 530f1e0e-5911-434e-bf20-edfd9778d7ad/530f1e0e-5911-434e-bf20-edfd9778d7ad.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 728.022727] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-143396e7-0ee5-4cf2-bc75-603d66ae8fcf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.047195] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 728.047195] env[66641]: value = "task-5145827" [ 728.047195] env[66641]: _type = "Task" [ 728.047195] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.055638] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145827, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.090237] env[66641]: DEBUG nova.scheduler.client.report [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 728.180870] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 728.181230] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-951b1a1e-178e-4db6-bb89-649d2f228c2c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.190435] env[66641]: DEBUG oslo_vmware.api [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 728.190435] env[66641]: value = "task-5145828" [ 728.190435] env[66641]: _type = "Task" [ 728.190435] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.201816] env[66641]: DEBUG oslo_vmware.api [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.257383] env[66641]: DEBUG oslo_concurrency.lockutils [req-7f1ca134-be49-43a8-a449-f1b77a926160 req-3ed8ea93-1176-4fce-84e3-6d5fae1ec3c1 service nova] Releasing lock "refresh_cache-530f1e0e-5911-434e-bf20-edfd9778d7ad" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 728.259082] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 728.300361] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 728.300707] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 728.300847] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 728.301119] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 728.301309] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 728.301465] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 728.301813] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.302062] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 728.302327] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 728.302554] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 728.302797] env[66641]: DEBUG nova.virt.hardware [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 728.303811] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0600d1b2-1276-4b37-8e07-e085d04049ad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.316167] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-719a2ffd-6358-4555-9d5c-13bb150ecc16 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.370024] env[66641]: DEBUG nova.compute.manager [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 728.370362] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.371574] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee197fe7-e3da-4eed-9f33-d79dd4961b04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.380707] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 728.381027] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e7367236-4840-4f85-9972-23dfeedd3441 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.464030] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 728.464030] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 728.464030] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleting the datastore file [datastore1] 695d1824-cf34-46c1-b623-b474d2f1f8ef {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 728.464030] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f49079e-abc0-4e88-a4ad-9ca4d8418d84 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.475438] env[66641]: DEBUG oslo_vmware.api [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 728.475438] env[66641]: value = "task-5145830" [ 728.475438] env[66641]: _type = "Task" [ 728.475438] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.483513] env[66641]: DEBUG oslo_vmware.api [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.558924] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145827, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.598106] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.393s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 728.599111] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 728.603244] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.902s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 728.603462] env[66641]: DEBUG nova.objects.instance [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lazy-loading 'resources' on Instance uuid ac5bb5a2-e630-42f0-be0c-24f52be81367 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 728.703710] env[66641]: DEBUG oslo_vmware.api [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145828, 'name': PowerOffVM_Task, 'duration_secs': 0.244287} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.704468] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 728.704759] env[66641]: DEBUG nova.compute.manager [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 728.706239] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7bda7b-1173-4af8-a664-1f134b620dd1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.987923] env[66641]: DEBUG oslo_vmware.api [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.41227} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.988412] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 728.988731] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 728.989050] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.989339] env[66641]: INFO nova.compute.manager [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Took 0.62 seconds to destroy the instance on the hypervisor. [ 728.989678] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 728.989970] env[66641]: DEBUG nova.compute.manager [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 728.990189] env[66641]: DEBUG nova.network.neutron [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 728.990798] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 728.991226] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.060399] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145827, 'name': ReconfigVM_Task, 'duration_secs': 0.567739} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.061076] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Reconfigured VM instance instance-0000001a to attach disk [datastore2] 530f1e0e-5911-434e-bf20-edfd9778d7ad/530f1e0e-5911-434e-bf20-edfd9778d7ad.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 729.061867] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ccd39a9e-44ec-43b1-9d55-ad4946d7fd61 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.071281] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 729.071281] env[66641]: value = "task-5145831" [ 729.071281] env[66641]: _type = "Task" [ 729.071281] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.083925] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145831, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.086523] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.086974] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.110623] env[66641]: DEBUG nova.compute.utils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 729.117300] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 729.117300] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 729.117300] env[66641]: WARNING neutronclient.v2_0.client [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.117300] env[66641]: WARNING neutronclient.v2_0.client [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 729.117908] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 729.118632] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 729.222045] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2073a8cc-fabc-482a-928e-4b659c604595 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.060s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 729.392773] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2318abb-0db1-4470-a4cb-8c7426994edc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.408044] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fae909-6112-480d-887d-bcff1e2d25ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.451829] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Successfully updated port: 0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 729.458278] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f1ccb3-a5b1-46c7-ad0f-1c9c75c51453 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.460868] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 729.460868] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 729.467751] env[66641]: DEBUG nova.policy [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd1dfe5b347f4871b7d6c0fb6770e1b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8449c95dd0f74492929e4d1d492ac9d2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 729.475576] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a10c6eb-2e1b-4b45-aba7-2f6afb277798 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.493434] env[66641]: DEBUG nova.compute.provider_tree [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.587391] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145831, 'name': Rename_Task, 'duration_secs': 0.178023} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.587736] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 729.588014] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b41890c-87bf-4af4-87a6-b96147fc08e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.598166] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 729.598166] env[66641]: value = "task-5145832" [ 729.598166] env[66641]: _type = "Task" [ 729.598166] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.612295] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145832, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.616213] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 729.820752] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Successfully created port: 2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 729.898432] env[66641]: DEBUG nova.network.neutron [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 729.964727] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.964917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquired lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 729.965140] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 729.966541] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 729.980840] env[66641]: DEBUG nova.compute.manager [req-dc76dbc9-f2d3-486b-a1c2-6e1024d508af req-6a767eb4-ae13-44e7-a9cf-681d895adc45 service nova] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Received event network-vif-deleted-932f4137-a36a-42bf-9b2f-ebfeaa078c3c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 730.000196] env[66641]: DEBUG nova.scheduler.client.report [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 730.120108] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145832, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.402273] env[66641]: INFO nova.compute.manager [-] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Took 1.41 seconds to deallocate network for instance. [ 730.471510] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 730.471510] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 730.476548] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 730.499187] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.506686] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.903s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 730.509530] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.523s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.510085] env[66641]: DEBUG nova.objects.instance [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lazy-loading 'resources' on Instance uuid eeb2556a-4e0c-43d7-83bd-942be5d5fdd2 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 730.532274] env[66641]: INFO nova.scheduler.client.report [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Deleted allocations for instance ac5bb5a2-e630-42f0-be0c-24f52be81367 [ 730.608602] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 730.608602] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 730.616608] env[66641]: DEBUG oslo_vmware.api [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145832, 'name': PowerOnVM_Task, 'duration_secs': 0.51618} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.617249] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 730.622056] env[66641]: INFO nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Took 12.29 seconds to spawn the instance on the hypervisor. [ 730.622056] env[66641]: DEBUG nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 730.622448] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe8ccfb-d85b-4ce7-b7f2-15b50fb9bae5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.627393] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 730.664363] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 730.664688] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 730.664854] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 730.665042] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 730.665216] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 730.665617] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 730.665617] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.665806] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 730.665976] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 730.666154] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 730.666324] env[66641]: DEBUG nova.virt.hardware [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 730.667587] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17dd29ad-36c9-4360-950e-b349b05d8f80 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.677476] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87ad608-530c-4e79-b774-d275fc5c7372 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.914729] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.041310] env[66641]: DEBUG oslo_concurrency.lockutils [None req-93638a3c-e29b-47a3-a302-bb6fa8ff91b0 tempest-ServerPasswordTestJSON-731178450 tempest-ServerPasswordTestJSON-731178450-project-member] Lock "ac5bb5a2-e630-42f0-be0c-24f52be81367" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.441s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.111202] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 731.146920] env[66641]: INFO nova.compute.manager [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Took 21.96 seconds to build instance. [ 731.247022] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22434e40-e82f-47a3-ad9b-b7389b4884ee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.256018] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b42b8e11-b578-4eb5-89cd-c6054cff509b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.288893] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bc44d6-9557-44ad-92e7-98a4449b8373 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.297433] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81b73cf-e005-4f64-b43f-9b9bb93d00d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.313387] env[66641]: DEBUG nova.compute.provider_tree [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.458294] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Successfully updated port: 2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 731.510954] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.511390] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.578894] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 731.579401] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 731.646857] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 731.648643] env[66641]: DEBUG oslo_concurrency.lockutils [None req-306cb112-cd2c-4c3d-af8b-cbced71418c6 tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.471s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 731.731136] env[66641]: DEBUG nova.network.neutron [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updating instance_info_cache with network_info: [{"id": "0b91b13d-bfce-46a3-acc4-217925393f9a", "address": "fa:16:3e:78:d6:e3", "network": {"id": "382d9765-4e05-4729-ba03-864a2036d224", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1929151786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "04f038652b90481f9629847d20f93f84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b91b13d-bf", "ovs_interfaceid": "0b91b13d-bfce-46a3-acc4-217925393f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 731.817576] env[66641]: DEBUG nova.scheduler.client.report [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 731.963583] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.963583] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 731.963583] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 732.234698] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Releasing lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 732.234975] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Instance network_info: |[{"id": "0b91b13d-bfce-46a3-acc4-217925393f9a", "address": "fa:16:3e:78:d6:e3", "network": {"id": "382d9765-4e05-4729-ba03-864a2036d224", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1929151786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "04f038652b90481f9629847d20f93f84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b91b13d-bf", "ovs_interfaceid": "0b91b13d-bfce-46a3-acc4-217925393f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 732.236077] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:d6:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2a75bb6e-6331-4429-b1b9-c968cc22b9c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b91b13d-bfce-46a3-acc4-217925393f9a', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 732.252571] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Creating folder: Project (04f038652b90481f9629847d20f93f84). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.252891] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fd8a8a98-137e-4a9e-937b-40a783883a75 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.265807] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Created folder: Project (04f038652b90481f9629847d20f93f84) in parent group-v1000566. [ 732.265963] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Creating folder: Instances. Parent ref: group-v1000645. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 732.266255] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43b94bba-bccd-4ea8-92d2-576727a3906f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.278713] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Created folder: Instances in parent group-v1000645. [ 732.279161] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 732.279424] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 732.279692] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-314fe6dd-6b06-4515-ab4e-c011787fe2fe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.304319] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 732.304319] env[66641]: value = "task-5145835" [ 732.304319] env[66641]: _type = "Task" [ 732.304319] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.315469] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145835, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.322553] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.813s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 732.325864] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.827s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 732.327673] env[66641]: INFO nova.compute.claims [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.365171] env[66641]: INFO nova.scheduler.client.report [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Deleted allocations for instance eeb2556a-4e0c-43d7-83bd-942be5d5fdd2 [ 732.383809] env[66641]: INFO nova.compute.manager [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Rebuilding instance [ 732.435617] env[66641]: DEBUG nova.compute.manager [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 732.436533] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdbf60b2-fa37-4f21-a5c8-03b93f6c8236 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.466249] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.467049] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.472788] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 732.815921] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145835, 'name': CreateVM_Task, 'duration_secs': 0.427363} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.815921] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 732.816686] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 732.817064] env[66641]: WARNING openstack [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 732.823441] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.823441] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 732.823441] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 732.823532] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-880c993b-cdc1-4910-9c16-2d4cfbf3603e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.829637] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 732.829637] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523db801-4275-94cb-04a7-f57835d757b0" [ 732.829637] env[66641]: _type = "Task" [ 732.829637] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.846690] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523db801-4275-94cb-04a7-f57835d757b0, 'name': SearchDatastore_Task, 'duration_secs': 0.01457} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.846932] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 732.847184] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.847437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.847594] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 732.847777] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.848071] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b94c6a0-257c-4883-a97b-d966bd472879 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.859414] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.859716] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 732.860756] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7e16e6c-efde-4f97-8530-c764b072b380 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.869381] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 732.869381] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52db3ad2-d008-734f-75b0-9064713d11fd" [ 732.869381] env[66641]: _type = "Task" [ 732.869381] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.879775] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52db3ad2-d008-734f-75b0-9064713d11fd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.880373] env[66641]: DEBUG oslo_concurrency.lockutils [None req-2343f775-4498-47ec-98bd-f3861c6dd709 tempest-TenantUsagesTestJSON-620064413 tempest-TenantUsagesTestJSON-620064413-project-member] Lock "eeb2556a-4e0c-43d7-83bd-942be5d5fdd2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.840s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.005635] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.006020] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.111890] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 733.112403] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 733.264803] env[66641]: DEBUG nova.network.neutron [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 733.323667] env[66641]: DEBUG nova.compute.manager [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Received event network-vif-plugged-0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 733.323994] env[66641]: DEBUG oslo_concurrency.lockutils [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] Acquiring lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.324326] env[66641]: DEBUG oslo_concurrency.lockutils [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.324592] env[66641]: DEBUG oslo_concurrency.lockutils [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 733.324849] env[66641]: DEBUG nova.compute.manager [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] No waiting events found dispatching network-vif-plugged-0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 733.325069] env[66641]: WARNING nova.compute.manager [req-4c67bca1-a5dc-47cd-8984-abb4ca34d561 req-d9ef06fb-cb9e-4655-941a-d0748941f422 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Received unexpected event network-vif-plugged-0b91b13d-bfce-46a3-acc4-217925393f9a for instance with vm_state building and task_state spawning. [ 733.382697] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52db3ad2-d008-734f-75b0-9064713d11fd, 'name': SearchDatastore_Task, 'duration_secs': 0.024893} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.383718] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66778543-026f-4d02-a407-232b37468cf4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.393186] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 733.393186] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c46e0b-59a7-cd8b-1c42-ae137d9ac4ff" [ 733.393186] env[66641]: _type = "Task" [ 733.393186] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.409122] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c46e0b-59a7-cd8b-1c42-ae137d9ac4ff, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.454899] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 733.455263] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e798a249-a6da-44de-8273-e2149b300f63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.464608] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 733.464608] env[66641]: value = "task-5145836" [ 733.464608] env[66641]: _type = "Task" [ 733.464608] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.476675] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145836, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.609981] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb7d1c9-e949-4a64-a804-d26bddce904e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.618753] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff858c5e-3db6-41f7-bd6f-d0e648ffc28e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.655885] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2946287-d644-4eb3-8a3f-8f02ff6b6333 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.667861] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf98a4f1-a227-4fff-b366-d4b40bdd452a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.682974] env[66641]: DEBUG nova.compute.provider_tree [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.753577] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "0f6055b9-f5b4-48ba-9589-0af212808be7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.754446] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.770131] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 733.770489] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Instance network_info: |[{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 733.771771] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e9:31:cc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a8b99a46-3e7f-4ef1-9e45-58e6cd17f210', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.778556] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 733.779070] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 733.779311] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dbcafa39-5a01-4df2-943b-204806362eb2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.803895] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.803895] env[66641]: value = "task-5145837" [ 733.803895] env[66641]: _type = "Task" [ 733.803895] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.812579] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145837, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.862316] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 733.862461] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 733.910166] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c46e0b-59a7-cd8b-1c42-ae137d9ac4ff, 'name': SearchDatastore_Task, 'duration_secs': 0.033426} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.910470] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 733.910732] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] b9d032da-031e-42e0-86e2-95254c1ceac1/b9d032da-031e-42e0-86e2-95254c1ceac1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 733.911164] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cde29371-6ddb-4067-a369-35d63ca1cfbf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.919757] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 733.919757] env[66641]: value = "task-5145838" [ 733.919757] env[66641]: _type = "Task" [ 733.919757] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.929680] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145838, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.978970] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145836, 'name': PowerOffVM_Task, 'duration_secs': 0.138924} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.979304] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 733.979537] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 733.980360] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a9b67b-9e7c-4feb-992b-2f593632f7f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.988493] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 733.988759] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b5d44d65-f5a9-4073-b5a6-b08da68961fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.021817] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 734.022067] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 734.022232] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Deleting the datastore file [datastore2] d03f39a4-532b-439a-9055-19fc1e769fff {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 734.022509] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b670dafe-48d5-4066-b3ea-13d82bf78190 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.033496] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 734.033496] env[66641]: value = "task-5145840" [ 734.033496] env[66641]: _type = "Task" [ 734.033496] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.044394] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145840, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.188260] env[66641]: DEBUG nova.scheduler.client.report [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 734.223292] env[66641]: DEBUG nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Received event network-vif-deleted-4c64a4dc-a032-4361-b383-ea4b1d7c3728 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 734.223292] env[66641]: DEBUG nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Received event network-vif-plugged-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 734.223494] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.224424] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.224974] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.225592] env[66641]: DEBUG nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] No waiting events found dispatching network-vif-plugged-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 734.227718] env[66641]: WARNING nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Received unexpected event network-vif-plugged-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a for instance with vm_state building and task_state spawning. [ 734.227718] env[66641]: DEBUG nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Received event network-changed-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 734.227718] env[66641]: DEBUG nova.compute.manager [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Refreshing instance network info cache due to event network-changed-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 734.227718] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Acquiring lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.227718] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Acquired lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.227943] env[66641]: DEBUG nova.network.neutron [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Refreshing network info cache for port 2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 734.256912] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 734.316417] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145837, 'name': CreateVM_Task, 'duration_secs': 0.392073} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.317370] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 734.318608] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.319187] env[66641]: WARNING openstack [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.325745] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.326145] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.326704] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 734.327537] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7eb9f001-0a27-48f9-8b53-fcf4160f2f4d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.335198] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 734.335198] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520af96e-a15c-9972-9242-cf2f98b00c3e" [ 734.335198] env[66641]: _type = "Task" [ 734.335198] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.346598] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520af96e-a15c-9972-9242-cf2f98b00c3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.365680] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 734.434452] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145838, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.548032] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145840, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.217025} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.548032] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 734.548032] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 734.548032] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.574178] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.575454] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.575979] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.575979] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.576200] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.581329] env[66641]: INFO nova.compute.manager [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Terminating instance [ 734.681184] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.681437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.694020] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 734.694553] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 734.697274] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.783s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 734.697478] env[66641]: DEBUG nova.objects.instance [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lazy-loading 'resources' on Instance uuid 695d1824-cf34-46c1-b623-b474d2f1f8ef {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 734.730757] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.731601] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 734.788725] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.848293] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520af96e-a15c-9972-9242-cf2f98b00c3e, 'name': SearchDatastore_Task, 'duration_secs': 0.069566} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.848493] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 734.848713] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 734.848990] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.849192] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 734.849422] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 734.849671] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e711fcdf-b23d-463e-92c1-6e7a4b584001 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.861440] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 734.861621] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 734.862394] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd5388d1-2ba1-43d0-bc94-6996e962e6da {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.868456] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 734.868456] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52268d17-1cc9-0278-2418-0c4bced4ff1b" [ 734.868456] env[66641]: _type = "Task" [ 734.868456] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.880459] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52268d17-1cc9-0278-2418-0c4bced4ff1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.887478] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 734.932272] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145838, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679874} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.932621] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] b9d032da-031e-42e0-86e2-95254c1ceac1/b9d032da-031e-42e0-86e2-95254c1ceac1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 734.932940] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.933341] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d16d06ec-28ef-41a8-9851-31a8944b06da {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.942122] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 734.942122] env[66641]: value = "task-5145841" [ 734.942122] env[66641]: _type = "Task" [ 734.942122] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.951241] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145841, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.970269] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 734.970628] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.082877] env[66641]: DEBUG nova.compute.manager [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 735.083317] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 735.084913] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e39cba47-ea64-4db5-99d6-e5ca09c36531 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.099114] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 735.099577] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-468eef63-b349-46fb-a470-48d1d0efe130 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.185301] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 735.191554] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 735.191640] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 735.192419] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleting the datastore file [datastore2] 3619be34-19ed-4d3d-b3ee-573126dcefbb {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 735.193953] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92b8abf6-d474-4f31-a493-ae61cefefcf7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.206758] env[66641]: DEBUG nova.compute.utils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 735.210401] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 735.213663] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 735.213663] env[66641]: WARNING neutronclient.v2_0.client [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.213663] env[66641]: WARNING neutronclient.v2_0.client [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 735.213663] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.214480] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.226849] env[66641]: DEBUG oslo_vmware.api [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 735.226849] env[66641]: value = "task-5145843" [ 735.226849] env[66641]: _type = "Task" [ 735.226849] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.241643] env[66641]: DEBUG oslo_vmware.api [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.338284] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.338284] env[66641]: WARNING openstack [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.385390] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52268d17-1cc9-0278-2418-0c4bced4ff1b, 'name': SearchDatastore_Task, 'duration_secs': 0.02134} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.389168] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9712e5e-ca29-471d-b4a9-37bf06496fbc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.395417] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 735.395417] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f5c7cd-7e42-8230-ed20-883a6da40677" [ 735.395417] env[66641]: _type = "Task" [ 735.395417] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.410050] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f5c7cd-7e42-8230-ed20-883a6da40677, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.463045] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145841, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.218981} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.463045] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.463426] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6d470b5-ecf7-47ea-9ea9-c53d79f00139 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.494919] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Reconfiguring VM instance instance-0000001b to attach disk [datastore1] b9d032da-031e-42e0-86e2-95254c1ceac1/b9d032da-031e-42e0-86e2-95254c1ceac1.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.498607] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-24b686bf-3f34-47bc-afc4-f8d2e681edc6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.521973] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 735.521973] env[66641]: value = "task-5145844" [ 735.521973] env[66641]: _type = "Task" [ 735.521973] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.531777] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145844, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.602678] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 735.604712] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.605236] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 735.605236] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 735.605236] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 735.605236] env[66641]: DEBUG nova.virt.hardware [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 735.605352] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d70df51-0b5c-45be-b1e7-85149312ca6d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.617532] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bc9ad62-06a8-448c-9258-8836b9e1515f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.641806] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 735.649019] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 735.651704] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 735.652237] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d3f34000-7a08-482b-b43c-3eb15f9b8341 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.672230] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 735.672230] env[66641]: value = "task-5145845" [ 735.672230] env[66641]: _type = "Task" [ 735.672230] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.682527] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145845, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.684594] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f44d2009-4582-40f3-b390-4003da3c8992 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.692263] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd21ffe5-e8b3-4916-9467-5d64533c2d7b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.731040] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 735.735042] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.739474] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4e1002-7b9b-4a91-b307-96c8bf343407 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.754289] env[66641]: DEBUG oslo_vmware.api [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.470891} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.757618] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 735.757854] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 735.758014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 735.758298] env[66641]: INFO nova.compute.manager [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Took 0.68 seconds to destroy the instance on the hypervisor. [ 735.758662] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 735.759061] env[66641]: DEBUG nova.compute.manager [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 735.759226] env[66641]: DEBUG nova.network.neutron [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 735.760049] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 735.760534] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 735.774341] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e82b16-94f3-467f-9b2a-a9e97ebf150b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.794551] env[66641]: DEBUG nova.compute.provider_tree [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.800602] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "05882781-78be-4568-95f4-2fccc4cf4dfe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 735.800778] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 735.912986] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f5c7cd-7e42-8230-ed20-883a6da40677, 'name': SearchDatastore_Task, 'duration_secs': 0.040966} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.913310] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 735.913578] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 735.915148] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2c973678-d8c0-47f7-84da-d37480ec4e66 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.923367] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 735.923367] env[66641]: value = "task-5145846" [ 735.923367] env[66641]: _type = "Task" [ 735.923367] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.932923] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145846, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.941995] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51cda0a9-e30e-4b3c-8211-1f7f0af165f4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.949790] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Suspending the VM {{(pid=66641) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 735.950088] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0c9dce21-eea0-40ad-98c3-dae7bfbb6064 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.958582] env[66641]: DEBUG oslo_vmware.api [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] Waiting for the task: (returnval){ [ 735.958582] env[66641]: value = "task-5145847" [ 735.958582] env[66641]: _type = "Task" [ 735.958582] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.967723] env[66641]: DEBUG oslo_vmware.api [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] Task: {'id': task-5145847, 'name': SuspendVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.031715] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145844, 'name': ReconfigVM_Task, 'duration_secs': 0.389505} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.031994] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Reconfigured VM instance instance-0000001b to attach disk [datastore1] b9d032da-031e-42e0-86e2-95254c1ceac1/b9d032da-031e-42e0-86e2-95254c1ceac1.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 736.032848] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1508a24a-2bbd-47ee-96e8-11a0fdbf60a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.041706] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 736.041706] env[66641]: value = "task-5145848" [ 736.041706] env[66641]: _type = "Task" [ 736.041706] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.055848] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145848, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.186026] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145845, 'name': CreateVM_Task, 'duration_secs': 0.350319} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.186026] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 736.186026] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.186026] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.186026] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 736.186026] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-051b64fa-4a99-4b3a-9c0e-857a35e8ca5a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.201162] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 736.201162] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52315599-b7b9-8b08-275a-436788bfde44" [ 736.201162] env[66641]: _type = "Task" [ 736.201162] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.213824] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52315599-b7b9-8b08-275a-436788bfde44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.239680] env[66641]: DEBUG nova.policy [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '25bd83aec32d4f60a32ce3538b84a664', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ecf362f0f454995adbf560fa17e1dab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 736.303047] env[66641]: DEBUG nova.scheduler.client.report [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 736.305654] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 736.346525] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 736.346829] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 736.439408] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145846, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.450228] env[66641]: DEBUG nova.network.neutron [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updated VIF entry in instance network info cache for port 2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 736.451330] env[66641]: DEBUG nova.network.neutron [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 736.474811] env[66641]: DEBUG oslo_vmware.api [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] Task: {'id': task-5145847, 'name': SuspendVM_Task} progress is 54%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.554864] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145848, 'name': Rename_Task, 'duration_secs': 0.180787} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.554978] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 736.555380] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60deb28b-0a5e-4810-878a-b2501f01219d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.566376] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 736.566376] env[66641]: value = "task-5145849" [ 736.566376] env[66641]: _type = "Task" [ 736.566376] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.577434] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145849, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.717283] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52315599-b7b9-8b08-275a-436788bfde44, 'name': SearchDatastore_Task, 'duration_secs': 0.038076} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.717283] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.717283] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 736.717995] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.717995] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 736.717995] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 736.719826] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2dab1d7d-af0d-4516-87bb-4144c174e6f7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.736610] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 736.736610] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 736.737174] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1757980-f9fb-409d-9d0a-47b36c7dcf14 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.748298] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 736.748298] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523f5a70-58ad-3606-e712-26f86e39420b" [ 736.748298] env[66641]: _type = "Task" [ 736.748298] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.753394] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 736.768940] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523f5a70-58ad-3606-e712-26f86e39420b, 'name': SearchDatastore_Task, 'duration_secs': 0.014175} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.770449] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4f81f22-780c-4c47-bf9c-62d01fda40d8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.778329] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 736.778329] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5278adbd-e7d8-8690-d75e-0b1a4f84b3b9" [ 736.778329] env[66641]: _type = "Task" [ 736.778329] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.787714] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 736.788815] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 736.788815] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 736.788815] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 736.788815] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 736.788815] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 736.789345] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.789345] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 736.789345] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 736.789440] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 736.789561] env[66641]: DEBUG nova.virt.hardware [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 736.790555] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce2ce08-e432-43fb-b243-c5b41b1cc99d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.796306] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5278adbd-e7d8-8690-d75e-0b1a4f84b3b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.802484] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a953c61f-d6a3-4db0-9525-b1fbddd5f3fa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.818085] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.121s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 736.826821] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.179s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 736.827559] env[66641]: INFO nova.compute.claims [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.846746] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 736.857711] env[66641]: INFO nova.scheduler.client.report [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleted allocations for instance 695d1824-cf34-46c1-b623-b474d2f1f8ef [ 736.874213] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Successfully created port: ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 736.938097] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145846, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.830516} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.938822] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 736.939070] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 736.939221] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-439f0dec-a8db-4f1a-b567-2f027616b5b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.949700] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 736.949700] env[66641]: value = "task-5145850" [ 736.949700] env[66641]: _type = "Task" [ 736.949700] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.953791] env[66641]: DEBUG oslo_concurrency.lockutils [req-cacbaff3-5597-46d7-8713-857355983000 req-505a4a8b-2250-4432-81c2-7d775f9d0b71 service nova] Releasing lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 736.964263] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145850, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.974629] env[66641]: DEBUG oslo_vmware.api [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] Task: {'id': task-5145847, 'name': SuspendVM_Task, 'duration_secs': 0.967047} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.974629] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Suspended the VM {{(pid=66641) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 736.974629] env[66641]: DEBUG nova.compute.manager [None req-2eb1ef5c-2e6a-4983-8a2d-266586f6c057 tempest-ServersAdminNegativeTestJSON-1348333929 tempest-ServersAdminNegativeTestJSON-1348333929-project-admin] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 736.975181] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfcca330-9e3a-4145-a3cc-227efb43db2e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.082476] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145849, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.293508] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5278adbd-e7d8-8690-d75e-0b1a4f84b3b9, 'name': SearchDatastore_Task, 'duration_secs': 0.014777} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.294115] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 737.294115] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 737.294547] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dbc0caba-3d6d-4c28-a331-5294ba8e37a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.302878] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 737.302878] env[66641]: value = "task-5145851" [ 737.302878] env[66641]: _type = "Task" [ 737.302878] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.312583] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145851, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.371025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e75c4b-efb6-4e5f-ac49-0d3df2a43329 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "695d1824-cf34-46c1-b623-b474d2f1f8ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.511s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 737.461397] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145850, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081742} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.462616] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 737.462698] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-399797aa-7d26-4cd8-b630-11efe528d771 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.492475] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 737.494968] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5ffcbde-a595-42e2-91be-aa0fe35a0b7a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.523176] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 737.523176] env[66641]: value = "task-5145852" [ 737.523176] env[66641]: _type = "Task" [ 737.523176] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.534934] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145852, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.585265] env[66641]: DEBUG oslo_vmware.api [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145849, 'name': PowerOnVM_Task, 'duration_secs': 0.613782} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.585265] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 737.585265] env[66641]: INFO nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Took 9.33 seconds to spawn the instance on the hypervisor. [ 737.585265] env[66641]: DEBUG nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 737.586206] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea99fbfd-9ccf-49cf-a680-e5cf16af8544 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.630830] env[66641]: DEBUG nova.network.neutron [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 737.726557] env[66641]: DEBUG nova.compute.manager [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Received event network-changed-0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 737.726821] env[66641]: DEBUG nova.compute.manager [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Refreshing instance network info cache due to event network-changed-0b91b13d-bfce-46a3-acc4-217925393f9a. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 737.727236] env[66641]: DEBUG oslo_concurrency.lockutils [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Acquiring lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.727517] env[66641]: DEBUG oslo_concurrency.lockutils [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Acquired lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 737.727659] env[66641]: DEBUG nova.network.neutron [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Refreshing network info cache for port 0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 737.815602] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145851, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.036482] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145852, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.111675] env[66641]: INFO nova.compute.manager [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Took 23.91 seconds to build instance. [ 738.127871] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689d8be7-97e7-40c2-8544-e071f1a3642b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.136430] env[66641]: INFO nova.compute.manager [-] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Took 2.38 seconds to deallocate network for instance. [ 738.144698] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a87b57-0aff-4bd7-aeb2-29af77bb2b6b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.184886] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160d4f2e-e090-478e-83f3-3fa4880e7280 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.194418] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9916079b-9fa7-4fd6-89c9-6e6599de2625 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.214035] env[66641]: DEBUG nova.compute.provider_tree [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 738.234498] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 738.234898] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 738.317321] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145851, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.60345} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.317503] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 738.317713] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.318315] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b21635fa-d145-4f38-8e51-23648843ccff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.326598] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 738.326598] env[66641]: value = "task-5145853" [ 738.326598] env[66641]: _type = "Task" [ 738.326598] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.336860] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145853, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.538463] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145852, 'name': ReconfigVM_Task, 'duration_secs': 0.679895} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.538463] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 738.538463] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3856694-ad02-4329-93b9-57904abb34ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.546890] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 738.546890] env[66641]: value = "task-5145854" [ 738.546890] env[66641]: _type = "Task" [ 738.546890] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.557944] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145854, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.614680] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dc543dd9-e519-4578-8935-259bd6480ea5 tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.425s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 738.655172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 738.717336] env[66641]: DEBUG nova.scheduler.client.report [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 738.804977] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Successfully updated port: ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 738.838726] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145853, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260272} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.843542] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.845310] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce5f428c-d7ab-41d0-b85b-4816328814f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.869982] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Reconfiguring VM instance instance-00000019 to attach disk [datastore1] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.869982] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c65679f7-519a-4709-bc24-cb8b72e83d22 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.894574] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 738.894574] env[66641]: value = "task-5145855" [ 738.894574] env[66641]: _type = "Task" [ 738.894574] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.904074] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145855, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.058159] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145854, 'name': Rename_Task, 'duration_secs': 0.284107} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.058482] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 739.058742] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-076e2018-a912-4528-aab5-0aacaccd0386 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.065639] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 739.065639] env[66641]: value = "task-5145856" [ 739.065639] env[66641]: _type = "Task" [ 739.065639] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.078817] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145856, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.222812] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.397s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 739.223735] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 739.227526] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.439s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 739.229187] env[66641]: INFO nova.compute.claims [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.281998] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.282490] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.312452] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.312671] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 739.312894] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 739.405530] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.578525] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145856, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.715500] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.715931] env[66641]: WARNING openstack [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.733687] env[66641]: DEBUG nova.compute.utils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 739.738335] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 739.738845] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 739.738887] env[66641]: WARNING neutronclient.v2_0.client [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.739288] env[66641]: WARNING neutronclient.v2_0.client [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 739.739990] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.741387] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.816821] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 739.816821] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 739.825877] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 739.906987] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145855, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.077886] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145856, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.236938] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 740.409029] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145855, 'name': ReconfigVM_Task, 'duration_secs': 1.249142} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.409583] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Reconfigured VM instance instance-00000019 to attach disk [datastore1] d03f39a4-532b-439a-9055-19fc1e769fff/d03f39a4-532b-439a-9055-19fc1e769fff.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.413533] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-06b0db7f-a8f6-43ba-8cdb-70f3aab2354d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.417819] env[66641]: DEBUG nova.network.neutron [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updated VIF entry in instance network info cache for port 0b91b13d-bfce-46a3-acc4-217925393f9a. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 740.420018] env[66641]: DEBUG nova.network.neutron [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updating instance_info_cache with network_info: [{"id": "0b91b13d-bfce-46a3-acc4-217925393f9a", "address": "fa:16:3e:78:d6:e3", "network": {"id": "382d9765-4e05-4729-ba03-864a2036d224", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1929151786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true}}], "meta": {"injected": false, "tenant_id": "04f038652b90481f9629847d20f93f84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b91b13d-bf", "ovs_interfaceid": "0b91b13d-bfce-46a3-acc4-217925393f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 740.424335] env[66641]: DEBUG nova.policy [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dfb62f2b801a4c7a899ac65c05c382f1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6624f269bb8840e98c10259f006ce962', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 740.429564] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 740.429564] env[66641]: value = "task-5145857" [ 740.429564] env[66641]: _type = "Task" [ 740.429564] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.445203] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145857, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.506609] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 740.507587] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 740.579232] env[66641]: DEBUG oslo_vmware.api [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145856, 'name': PowerOnVM_Task, 'duration_secs': 1.297993} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.579563] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 740.579768] env[66641]: INFO nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Took 9.95 seconds to spawn the instance on the hypervisor. [ 740.580799] env[66641]: DEBUG nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 740.580799] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9928babe-2041-47bc-bc62-c1674e98b225 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.587716] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47b2ee7-3dd7-45eb-9add-0f8c4a66bfef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.605708] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-906be670-dcf1-4e29-a603-0b568ccad24e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.654865] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5063b840-78f2-4e71-aea5-217aedadd462 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.663775] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9121cc-8862-4649-be7f-d25a950ade5c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.682665] env[66641]: DEBUG nova.compute.provider_tree [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.849106] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Successfully created port: a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 740.930726] env[66641]: DEBUG oslo_concurrency.lockutils [req-b789ce4c-9db0-452a-bdd8-8072929ad718 req-a50f777b-0c70-4509-b63c-487a5238f834 service nova] Releasing lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 740.944649] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145857, 'name': Rename_Task, 'duration_secs': 0.222608} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.944962] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 740.945220] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8c44cd03-a1b2-4c97-b3c6-553282d5c6b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.952917] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 740.952917] env[66641]: value = "task-5145858" [ 740.952917] env[66641]: _type = "Task" [ 740.952917] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.962027] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145858, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.111971] env[66641]: INFO nova.compute.manager [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Took 25.56 seconds to build instance. [ 741.188958] env[66641]: DEBUG nova.scheduler.client.report [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 741.250976] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 741.268875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "dad5ed23-71a1-4b55-856f-2484f8e62708" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.269235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.291636] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:51:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ed028205-0b7e-4e53-9191-4b8a2662b934',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-192737414',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 741.291887] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 741.292128] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 741.292419] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 741.292596] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 741.292740] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 741.293048] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.293342] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 741.293543] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 741.293692] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 741.293837] env[66641]: DEBUG nova.virt.hardware [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 741.295535] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cf51d9-2228-46b0-9a87-d5ad987e7f7e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.306487] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934307d4-dbcb-4c9a-9827-98a12c4f41ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.468206] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145858, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.570604] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 741.571083] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 741.615640] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b2300575-ed70-4c05-a002-d41cf32d1646 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.075s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.700610] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 741.701146] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 741.703945] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.817s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.705564] env[66641]: INFO nova.compute.claims [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.774181] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 741.883101] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 741.883101] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 741.967738] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145858, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.096223] env[66641]: DEBUG nova.network.neutron [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Updating instance_info_cache with network_info: [{"id": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "address": "fa:16:3e:dd:26:d5", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1fff4b-fc", "ovs_interfaceid": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 742.173927] env[66641]: DEBUG nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Received event network-vif-deleted-5a4d98a2-57b2-40e0-9616-a27818d66651 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 742.174028] env[66641]: DEBUG nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Received event network-vif-plugged-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 742.174434] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Acquiring lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.174740] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 742.175035] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 742.175335] env[66641]: DEBUG nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] No waiting events found dispatching network-vif-plugged-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 742.175589] env[66641]: WARNING nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Received unexpected event network-vif-plugged-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 for instance with vm_state building and task_state spawning. [ 742.176568] env[66641]: DEBUG nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Received event network-changed-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 742.176747] env[66641]: DEBUG nova.compute.manager [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Refreshing instance network info cache due to event network-changed-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 742.176954] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Acquiring lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.212018] env[66641]: DEBUG nova.compute.utils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 742.215619] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 742.215850] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 742.216270] env[66641]: WARNING neutronclient.v2_0.client [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.216527] env[66641]: WARNING neutronclient.v2_0.client [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 742.217280] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 742.217994] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 742.312294] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.386166] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 742.466766] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145858, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.492090] env[66641]: DEBUG nova.policy [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1eb90a1845bb4360816715d1e0f51b5e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '417e1e40aca447229001fa725c82e9ca', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 742.537813] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Successfully updated port: a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 742.553802] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc85925-1e12-44a6-b623-e43705c1f82b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.565710] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94265740-1322-45ed-9339-bb71975c7c05 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.605570] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 742.605874] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Instance network_info: |[{"id": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "address": "fa:16:3e:dd:26:d5", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1fff4b-fc", "ovs_interfaceid": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 742.606647] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Acquired lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 742.606846] env[66641]: DEBUG nova.network.neutron [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Refreshing network info cache for port ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 742.609022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:26:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 742.616020] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating folder: Project (7ecf362f0f454995adbf560fa17e1dab). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.616860] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c706822e-1bcd-4e6b-8539-21375630d1d6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.620843] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89161b7d-1881-4a5e-84d7-f6e5c6b41190 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.629682] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f724d961-317e-4590-89f4-1c73a891185b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.636682] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Created folder: Project (7ecf362f0f454995adbf560fa17e1dab) in parent group-v1000566. [ 742.636915] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating folder: Instances. Parent ref: group-v1000650. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 742.638080] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c7e45e3-c91f-4cb2-aad7-76b5ad5ec314 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.650443] env[66641]: DEBUG nova.compute.provider_tree [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.662514] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Created folder: Instances in parent group-v1000650. [ 742.663146] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 742.663829] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 742.664010] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-23651c37-6f86-4cc8-ac0c-4a4cd918cc6a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.688916] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 742.688916] env[66641]: value = "task-5145861" [ 742.688916] env[66641]: _type = "Task" [ 742.688916] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.698890] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145861, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.729077] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 742.903526] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Successfully created port: 298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 742.914198] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 742.970659] env[66641]: DEBUG oslo_vmware.api [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145858, 'name': PowerOnVM_Task, 'duration_secs': 1.876774} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.974910] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 742.974910] env[66641]: DEBUG nova.compute.manager [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 742.974910] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e529117-1b62-417a-92b1-6e8b7a275117 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.041568] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.041807] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 743.042043] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 743.122353] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.122747] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.157025] env[66641]: DEBUG nova.scheduler.client.report [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 743.202765] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145861, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.500177] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 743.545560] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 743.545560] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 743.555162] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 743.661421] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.957s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 743.661693] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 743.665804] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.931s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 743.667258] env[66641]: INFO nova.compute.claims [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.705302] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145861, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.742956] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 743.774728] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 743.774971] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 743.775129] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 743.775308] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 743.775447] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 743.775585] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 743.775796] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.775951] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 743.776113] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 743.776279] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 743.776447] env[66641]: DEBUG nova.virt.hardware [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 743.777348] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d239be3e-59c5-46d2-abbb-8fe829fa2b74 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.785723] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d9f3cd-790c-4c65-b249-a008f22f9e1f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.173656] env[66641]: DEBUG nova.compute.utils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 744.180572] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 744.180572] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 744.180572] env[66641]: WARNING neutronclient.v2_0.client [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.181608] env[66641]: WARNING neutronclient.v2_0.client [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 744.182450] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.182951] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.202144] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145861, 'name': CreateVM_Task, 'duration_secs': 1.416754} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.205222] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 744.205222] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.205222] env[66641]: WARNING openstack [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.209256] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.209543] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.210224] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 744.210709] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3323cf85-4938-4629-b0bb-462484703772 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.216462] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 744.216462] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc0ba3-2369-3b92-bb94-8ade235238ed" [ 744.216462] env[66641]: _type = "Task" [ 744.216462] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.226275] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc0ba3-2369-3b92-bb94-8ade235238ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.375186] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.375656] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.535137] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Successfully updated port: 298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 744.551181] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 744.551712] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 744.681215] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 744.728411] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fc0ba3-2369-3b92-bb94-8ade235238ed, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.733027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 744.733027] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.733027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.733027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 744.733222] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.733222] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-39fbb8f1-7900-4d4d-8f61-2e3492d51c35 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.742787] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.743365] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 744.746362] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7569eba2-1c1c-4bdd-9fa3-2c92fb261b9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.752707] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 744.752707] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285934d-f91f-1cd8-7483-58bb960b9701" [ 744.752707] env[66641]: _type = "Task" [ 744.752707] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.764884] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285934d-f91f-1cd8-7483-58bb960b9701, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.977227] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f7a3bc-d733-40e8-8985-70f0ae8964d0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.987539] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e0bd6c0-d3db-4a27-afeb-422ca63b9f1d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.024231] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed57ecc2-da04-4f5d-be0d-2fe662f3c437 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.032776] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39714cce-d22d-4701-93bc-1650f13b2f97 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.037831] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.037982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 745.038070] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 745.053756] env[66641]: DEBUG nova.compute.provider_tree [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.268031] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285934d-f91f-1cd8-7483-58bb960b9701, 'name': SearchDatastore_Task, 'duration_secs': 0.010001} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.268031] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-03f6de4f-e766-46b4-bc93-437db197d12f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.274429] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 745.274429] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52818eb0-dccb-0f43-4fda-a00dccf97555" [ 745.274429] env[66641]: _type = "Task" [ 745.274429] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.286555] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52818eb0-dccb-0f43-4fda-a00dccf97555, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.446509] env[66641]: DEBUG nova.policy [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '25bd83aec32d4f60a32ce3538b84a664', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ecf362f0f454995adbf560fa17e1dab', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 745.543871] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.543871] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.549390] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 745.563127] env[66641]: DEBUG nova.scheduler.client.report [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 745.629711] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.629711] env[66641]: WARNING openstack [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.670555] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 745.670938] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 745.693324] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 745.725777] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 745.726142] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 745.726316] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 745.726500] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 745.726640] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 745.726778] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 745.726980] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.727163] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 745.727587] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 745.727587] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 745.727679] env[66641]: DEBUG nova.virt.hardware [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 745.728570] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54250d54-c3ef-4e28-8e5e-35b13e4c15cb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.739170] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22208c75-86b0-474e-a101-53a84cff1d07 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.784920] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52818eb0-dccb-0f43-4fda-a00dccf97555, 'name': SearchDatastore_Task, 'duration_secs': 0.012967} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.785230] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 745.785495] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a7575399-f72d-4d2c-b57b-f2f2f591c8a7/a7575399-f72d-4d2c-b57b-f2f2f591c8a7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 745.785769] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c41dead-45a5-4293-bf36-65a0243d1376 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.794345] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 745.794345] env[66641]: value = "task-5145862" [ 745.794345] env[66641]: _type = "Task" [ 745.794345] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.802911] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145862, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.955066] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "081457ae-e152-410c-bca7-4d43b95eee10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 745.955380] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 745.986738] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Successfully created port: ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 746.070693] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.405s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 746.071541] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 746.079027] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.229s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 746.079027] env[66641]: INFO nova.compute.claims [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.255397] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.257087] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.305517] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145862, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.400402] env[66641]: DEBUG nova.network.neutron [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Updated VIF entry in instance network info cache for port ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 746.401150] env[66641]: DEBUG nova.network.neutron [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Updating instance_info_cache with network_info: [{"id": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "address": "fa:16:3e:dd:26:d5", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapca1fff4b-fc", "ovs_interfaceid": "ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.438981] env[66641]: DEBUG nova.network.neutron [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 746.457714] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 746.595555] env[66641]: DEBUG nova.compute.utils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 746.598947] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 746.599259] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 746.599727] env[66641]: WARNING neutronclient.v2_0.client [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.600139] env[66641]: WARNING neutronclient.v2_0.client [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 746.600785] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.603496] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 746.806480] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145862, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.906272] env[66641]: DEBUG oslo_concurrency.lockutils [req-174273da-deb7-4a2d-91a2-ae06561efa70 req-9bd71cfe-e815-4c94-859a-8cd0b93a8918 service nova] Releasing lock "refresh_cache-a7575399-f72d-4d2c-b57b-f2f2f591c8a7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.945288] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 746.945645] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Instance network_info: |[{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 746.952016] env[66641]: DEBUG nova.policy [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '880080387dc04a88a4ec415e4d40fbed', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '400b5d331a5c48a0947adc3477fe65bb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 746.955485] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:05:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a44fab6d-1a8a-4a32-93c1-ebfd24e3d021', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.963319] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 746.966052] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 746.966296] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8028792d-8ef5-4e49-9fdb-6019189056cc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.991086] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 746.991424] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.002160] env[66641]: DEBUG nova.compute.manager [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Received event network-vif-plugged-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 747.002497] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Acquiring lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.002569] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 747.002700] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 747.002857] env[66641]: DEBUG nova.compute.manager [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] No waiting events found dispatching network-vif-plugged-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 747.003031] env[66641]: WARNING nova.compute.manager [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Received unexpected event network-vif-plugged-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 for instance with vm_state building and task_state spawning. [ 747.003207] env[66641]: DEBUG nova.compute.manager [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Received event network-changed-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 747.003362] env[66641]: DEBUG nova.compute.manager [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Refreshing instance network info cache due to event network-changed-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 747.003536] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.003662] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.003812] env[66641]: DEBUG nova.network.neutron [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Refreshing network info cache for port a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 747.011483] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.011483] env[66641]: value = "task-5145863" [ 747.011483] env[66641]: _type = "Task" [ 747.011483] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.024300] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145863, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.025476] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 747.099525] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 747.308026] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145862, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.367101] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96902e03-6775-4278-bb79-da6a58603c91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.375458] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e1aef8-997d-46fb-9d98-cf5f834aadcd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.416210] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a22c5a09-e116-4cc6-a22e-1519fc4f37cc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.425199] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d265da6-3f8c-45c7-81ca-25eb4b28cc86 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.439713] env[66641]: DEBUG nova.compute.provider_tree [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 747.508338] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 747.508852] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 747.529258] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145863, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.724868] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Successfully created port: 20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 747.804426] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Successfully updated port: ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 747.809884] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145862, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.708351} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.810647] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a7575399-f72d-4d2c-b57b-f2f2f591c8a7/a7575399-f72d-4d2c-b57b-f2f2f591c8a7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 747.811126] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.811539] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1bd0de5b-7977-4d3a-bfbc-5ab5dd8b4bd0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.822022] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 747.822022] env[66641]: value = "task-5145864" [ 747.822022] env[66641]: _type = "Task" [ 747.822022] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.831794] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.860864] env[66641]: DEBUG nova.compute.manager [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Received event network-changed-0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 747.862046] env[66641]: DEBUG nova.compute.manager [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Refreshing instance network info cache due to event network-changed-0b91b13d-bfce-46a3-acc4-217925393f9a. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 747.862046] env[66641]: DEBUG oslo_concurrency.lockutils [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Acquiring lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.862046] env[66641]: DEBUG oslo_concurrency.lockutils [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Acquired lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 747.862046] env[66641]: DEBUG nova.network.neutron [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Refreshing network info cache for port 0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 747.942987] env[66641]: DEBUG nova.scheduler.client.report [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 748.026949] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145863, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.066793] env[66641]: DEBUG nova.network.neutron [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Updating instance_info_cache with network_info: [{"id": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "address": "fa:16:3e:4f:a9:07", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298db8ec-67", "ovs_interfaceid": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 748.112519] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 748.142101] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 748.142255] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 748.142552] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 748.142620] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 748.142751] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 748.142895] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 748.143113] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 748.143628] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 748.143817] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 748.144180] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 748.144468] env[66641]: DEBUG nova.virt.hardware [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 748.145596] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eeb76b5-8294-44c0-b215-cc8ac2606e64 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.156068] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1cab1e8-0111-488d-8426-dd7da993c18f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.312013] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.312307] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 748.312473] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 748.338881] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082223} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.338881] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 748.338881] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fd29b2-516d-485d-9359-fc55d58cc253 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.367537] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Reconfiguring VM instance instance-0000001d to attach disk [datastore2] a7575399-f72d-4d2c-b57b-f2f2f591c8a7/a7575399-f72d-4d2c-b57b-f2f2f591c8a7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 748.368206] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.368773] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.376466] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5d7f048-cfd8-428e-b802-204985ac6e96 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.401152] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 748.401152] env[66641]: value = "task-5145865" [ 748.401152] env[66641]: _type = "Task" [ 748.401152] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.415180] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145865, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.452202] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.375s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 748.452202] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 748.455632] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.800s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 748.457079] env[66641]: DEBUG nova.objects.instance [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lazy-loading 'resources' on Instance uuid 3619be34-19ed-4d3d-b3ee-573126dcefbb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.529094] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145863, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.569602] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 748.570296] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Instance network_info: |[{"id": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "address": "fa:16:3e:4f:a9:07", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298db8ec-67", "ovs_interfaceid": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 748.571067] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:a9:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4b5c60ce-845e-4506-bc10-348461fece6d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '298db8ec-67aa-4b6a-8981-4f1a72602ef3', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.586134] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 748.587199] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 748.587677] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f5638f61-fbc4-4f4d-8592-9a66b4db791a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.623960] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.623960] env[66641]: value = "task-5145866" [ 748.623960] env[66641]: _type = "Task" [ 748.623960] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.634062] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145866, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.817233] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.817964] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.823955] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 748.916363] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145865, 'name': ReconfigVM_Task, 'duration_secs': 0.325573} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.916363] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Reconfigured VM instance instance-0000001d to attach disk [datastore2] a7575399-f72d-4d2c-b57b-f2f2f591c8a7/a7575399-f72d-4d2c-b57b-f2f2f591c8a7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.916363] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1dd56b5c-686f-4f0d-82c0-af4f10f110c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.922217] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 748.922217] env[66641]: value = "task-5145867" [ 748.922217] env[66641]: _type = "Task" [ 748.922217] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.930609] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145867, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.961325] env[66641]: DEBUG nova.compute.utils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 748.965442] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 748.965646] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 748.965955] env[66641]: WARNING neutronclient.v2_0.client [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.966285] env[66641]: WARNING neutronclient.v2_0.client [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 748.966850] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 748.967469] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 748.976868] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 749.028756] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.029458] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.044892] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145863, 'name': CreateVM_Task, 'duration_secs': 1.681134} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.044892] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 749.044892] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.044892] env[66641]: WARNING openstack [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.049354] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.049528] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.049847] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 749.053410] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-909cbbc5-3d40-400d-823f-227a140fdc53 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.064332] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 749.064332] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb71f0-cc98-b77b-d942-ed25689e559b" [ 749.064332] env[66641]: _type = "Task" [ 749.064332] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.076347] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb71f0-cc98-b77b-d942-ed25689e559b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.134833] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145866, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.308074] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4205dc9f-9f93-4ae5-89f5-09e7bac11b20 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.317871] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093d6cb6-2e3c-413d-bd0a-627831addd4d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.356513] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facc12e4-3a1c-459b-ba7b-c14448121d03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.365111] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef72f6a-3609-4ca7-8caf-f1c966bf2eec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.379609] env[66641]: DEBUG nova.compute.provider_tree [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 749.435453] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145867, 'name': Rename_Task, 'duration_secs': 0.305594} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.435453] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 749.435453] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c33b7a5-f11b-499e-8bae-0d447c3a1f3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.441511] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 749.441511] env[66641]: value = "task-5145868" [ 749.441511] env[66641]: _type = "Task" [ 749.441511] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.449668] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.477261] env[66641]: DEBUG nova.policy [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf590217db704af2acb3d078ce6d789f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17792883c6504d0f9f91fb6beb25b087', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 749.492857] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Successfully updated port: 20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 749.575875] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bb71f0-cc98-b77b-d942-ed25689e559b, 'name': SearchDatastore_Task, 'duration_secs': 0.010582} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.575969] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 749.576467] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 749.577510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.577510] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 749.577510] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.577510] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64fe76ee-07d7-4f52-8d5c-b48ab00571d8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.587602] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.587786] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 749.588577] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be6e5275-d429-435b-89f6-4a5d07884f60 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.597797] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 749.597797] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5227be84-b3b8-25a9-a4aa-79effc45b3ad" [ 749.597797] env[66641]: _type = "Task" [ 749.597797] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.606926] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5227be84-b3b8-25a9-a4aa-79effc45b3ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.638737] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145866, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.762204] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 749.762204] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 749.885943] env[66641]: DEBUG nova.scheduler.client.report [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 749.954347] env[66641]: DEBUG oslo_vmware.api [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145868, 'name': PowerOnVM_Task, 'duration_secs': 0.480947} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.954974] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 749.955251] env[66641]: INFO nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 13.20 seconds to spawn the instance on the hypervisor. [ 749.955839] env[66641]: DEBUG nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 749.956521] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bc1f5d1-9833-4482-be39-ce46df8197c8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.986846] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 750.000385] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.000385] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquired lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.000385] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 750.029864] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 750.030385] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 750.030466] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 750.030965] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 750.030965] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 750.030965] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 750.031166] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 750.031233] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 750.031411] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 750.031618] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 750.031792] env[66641]: DEBUG nova.virt.hardware [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 750.033094] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15825058-7bbc-46b9-ab24-c2a72cb2fe11 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.043350] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e0899d8-d981-4cc8-a7ca-d418c1a6f4f2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.105743] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Successfully created port: 923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 750.116497] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5227be84-b3b8-25a9-a4aa-79effc45b3ad, 'name': SearchDatastore_Task, 'duration_secs': 0.01109} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.118334] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9494007-6b5d-449f-a8c0-4972a40ca9ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.124519] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 750.124519] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e2d616-fca2-31cf-db1c-c4c70c2e14ce" [ 750.124519] env[66641]: _type = "Task" [ 750.124519] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.137486] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.137486] env[66641]: WARNING openstack [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.154461] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145866, 'name': CreateVM_Task, 'duration_secs': 1.413283} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.154721] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e2d616-fca2-31cf-db1c-c4c70c2e14ce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.154901] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 750.155728] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.156091] env[66641]: WARNING openstack [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.164832] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.164832] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.164832] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 750.165097] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a683473-ffec-4fa6-9099-6c336bfda46c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.171942] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 750.171942] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b99327-9187-7087-2515-371ee04e926e" [ 750.171942] env[66641]: _type = "Task" [ 750.171942] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.182138] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b99327-9187-7087-2515-371ee04e926e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.186939] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.187323] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.394974] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.939s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.397535] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.086s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 750.399185] env[66641]: INFO nova.compute.claims [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 750.422702] env[66641]: INFO nova.scheduler.client.report [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleted allocations for instance 3619be34-19ed-4d3d-b3ee-573126dcefbb [ 750.479492] env[66641]: INFO nova.compute.manager [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 20.00 seconds to build instance. [ 750.503615] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.504085] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.509279] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 750.535682] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.536147] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.638700] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e2d616-fca2-31cf-db1c-c4c70c2e14ce, 'name': SearchDatastore_Task, 'duration_secs': 0.03159} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.640471] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.640471] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 750.640471] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-861f58f0-2f3d-4291-801f-5d60fd1b6f03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.655708] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 750.655708] env[66641]: value = "task-5145869" [ 750.655708] env[66641]: _type = "Task" [ 750.655708] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.666964] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.691466] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b99327-9187-7087-2515-371ee04e926e, 'name': SearchDatastore_Task, 'duration_secs': 0.011666} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.691961] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 750.692439] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 750.692921] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 750.693046] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 750.693366] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 750.693812] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cd4fd99e-f6a9-4e22-8b00-858345118266 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.704916] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 750.705052] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 750.706233] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e48d3659-1ab5-4ecb-bc8d-20b2f6c585d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.716072] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 750.716072] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525533e2-9c91-c6ad-fc3d-6a7d55048d06" [ 750.716072] env[66641]: _type = "Task" [ 750.716072] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.728577] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525533e2-9c91-c6ad-fc3d-6a7d55048d06, 'name': SearchDatastore_Task, 'duration_secs': 0.010825} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.729060] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1730b5a8-b294-4fa2-8d91-9562b8223a6f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.735493] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 750.735493] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f10135-4f5b-f3cf-7084-2f69c053f6ca" [ 750.735493] env[66641]: _type = "Task" [ 750.735493] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.749037] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f10135-4f5b-f3cf-7084-2f69c053f6ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.758575] env[66641]: DEBUG nova.compute.manager [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Stashing vm_state: active {{(pid=66641) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 750.892085] env[66641]: DEBUG nova.network.neutron [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updated VIF entry in instance network info cache for port a44fab6d-1a8a-4a32-93c1-ebfd24e3d021. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 750.892474] env[66641]: DEBUG nova.network.neutron [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 750.926316] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 750.926765] env[66641]: WARNING openstack [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 750.938884] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bba4a975-4bd1-48f1-b7b8-fff451a4c82e tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "3619be34-19ed-4d3d-b3ee-573126dcefbb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.363s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 750.982089] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12dcb512-6c92-4ba9-9ab5-9b9d443c4b21 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.521s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.059819] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Received event network-vif-plugged-298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 751.059819] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquiring lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.059913] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.060408] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.060526] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] No waiting events found dispatching network-vif-plugged-298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 751.060783] env[66641]: WARNING nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Received unexpected event network-vif-plugged-298db8ec-67aa-4b6a-8981-4f1a72602ef3 for instance with vm_state building and task_state spawning. [ 751.061187] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Received event network-changed-298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 751.061254] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Refreshing instance network info cache due to event network-changed-298db8ec-67aa-4b6a-8981-4f1a72602ef3. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 751.062110] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquiring lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.062110] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquired lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.062297] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Refreshing network info cache for port 298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 751.154241] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.154568] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.154898] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.155172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.155370] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.161074] env[66641]: INFO nova.compute.manager [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Terminating instance [ 751.169774] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51305} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.170173] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.170462] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.170731] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-abc21167-b228-411f-8824-8f027b8b4e95 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.178855] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 751.178855] env[66641]: value = "task-5145870" [ 751.178855] env[66641]: _type = "Task" [ 751.178855] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.192878] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145870, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.248164] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f10135-4f5b-f3cf-7084-2f69c053f6ca, 'name': SearchDatastore_Task, 'duration_secs': 0.009376} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.248527] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.248844] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 0f6055b9-f5b4-48ba-9589-0af212808be7/0f6055b9-f5b4-48ba-9589-0af212808be7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 751.249288] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "d03f39a4-532b-439a-9055-19fc1e769fff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.249536] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.249792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "d03f39a4-532b-439a-9055-19fc1e769fff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.250025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 751.250258] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 751.251881] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbbe21cf-5d2b-4845-aaf1-f626e373066d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.254441] env[66641]: INFO nova.compute.manager [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Terminating instance [ 751.262395] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 751.262395] env[66641]: value = "task-5145871" [ 751.262395] env[66641]: _type = "Task" [ 751.262395] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.281382] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145871, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.283066] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.283480] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.291235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 751.321892] env[66641]: DEBUG nova.network.neutron [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Updating instance_info_cache with network_info: [{"id": "ee7db498-4256-4265-9b73-6b6b2afa057c", "address": "fa:16:3e:f9:cb:a3", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.185", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7db498-42", "ovs_interfaceid": "ee7db498-4256-4265-9b73-6b6b2afa057c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 751.395614] env[66641]: DEBUG oslo_concurrency.lockutils [req-eba6f84a-58e8-44bb-b7d1-74fba80b042b req-f4a66284-5465-40ef-af69-8e6fd6fc8c44 service nova] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.566353] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 751.568605] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 751.667693] env[66641]: DEBUG nova.compute.manager [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 751.667966] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 751.668981] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af5a9b8-66ba-458d-90ab-781b28fa878a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.680238] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 751.682682] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60371fb8-6b78-4f25-9e5e-227875640c08 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.695913] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145870, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07687} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.695913] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.696078] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28114fd0-2059-4764-8db5-10ee01ea73f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.701617] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 751.701617] env[66641]: value = "task-5145872" [ 751.701617] env[66641]: _type = "Task" [ 751.701617] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.729888] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.733479] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-53054879-c8a7-419f-a9ed-6c62f1f0ca2d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.753130] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.759331] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "refresh_cache-d03f39a4-532b-439a-9055-19fc1e769fff" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.759507] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquired lock "refresh_cache-d03f39a4-532b-439a-9055-19fc1e769fff" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 751.759684] env[66641]: DEBUG nova.network.neutron [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 751.762941] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 751.762941] env[66641]: value = "task-5145873" [ 751.762941] env[66641]: _type = "Task" [ 751.762941] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.781176] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145871, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487939} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.781817] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.782127] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 0f6055b9-f5b4-48ba-9589-0af212808be7/0f6055b9-f5b4-48ba-9589-0af212808be7.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 751.782489] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.782617] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9c8ee0b-5d99-4091-af42-29970616496b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.790780] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 751.790780] env[66641]: value = "task-5145874" [ 751.790780] env[66641]: _type = "Task" [ 751.790780] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.805091] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145874, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.822903] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 751.823550] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Instance network_info: |[{"id": "ee7db498-4256-4265-9b73-6b6b2afa057c", "address": "fa:16:3e:f9:cb:a3", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.185", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7db498-42", "ovs_interfaceid": "ee7db498-4256-4265-9b73-6b6b2afa057c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 751.824378] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:cb:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '193994c7-8e1b-4f25-a4a4-d0563845eb28', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ee7db498-4256-4265-9b73-6b6b2afa057c', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.832761] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 751.834681] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.835488] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e725f4e1-61e1-4565-870e-cf14b5c02bee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.838358] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f1b3223-3ead-49b6-b2f9-7ab93e952256 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.862687] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16245b72-6d87-492f-95fc-b34ab30f679a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.869240] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.869240] env[66641]: value = "task-5145875" [ 751.869240] env[66641]: _type = "Task" [ 751.869240] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.903251] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7889e68-d701-4c2a-a184-04b053ed1cdf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.909690] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145875, 'name': CreateVM_Task} progress is 15%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.917826] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c306d9b6-e2a1-4347-bab4-383cc4eef147 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.937382] env[66641]: DEBUG nova.compute.provider_tree [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.987572] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Successfully updated port: 923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 752.025675] env[66641]: DEBUG nova.network.neutron [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updated VIF entry in instance network info cache for port 0b91b13d-bfce-46a3-acc4-217925393f9a. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 752.026108] env[66641]: DEBUG nova.network.neutron [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updating instance_info_cache with network_info: [{"id": "0b91b13d-bfce-46a3-acc4-217925393f9a", "address": "fa:16:3e:78:d6:e3", "network": {"id": "382d9765-4e05-4729-ba03-864a2036d224", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-1929151786-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.209", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04f038652b90481f9629847d20f93f84", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2a75bb6e-6331-4429-b1b9-c968cc22b9c9", "external-id": "nsx-vlan-transportzone-244", "segmentation_id": 244, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b91b13d-bf", "ovs_interfaceid": "0b91b13d-bfce-46a3-acc4-217925393f9a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.219140] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145872, 'name': PowerOffVM_Task, 'duration_secs': 0.263367} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.219336] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 752.219428] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 752.219724] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a97f689-6409-446f-a687-cb64ec06c314 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.263807] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.265516] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.274834] env[66641]: WARNING openstack [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.275146] env[66641]: WARNING openstack [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.281028] env[66641]: DEBUG nova.network.neutron [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 752.295831] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145873, 'name': ReconfigVM_Task, 'duration_secs': 0.416602} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.306127] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.308433] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-baa79b37-7a69-4fa9-8464-6632e332e311 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.315771] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 752.316349] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 752.316497] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleting the datastore file [datastore2] fb1bf80c-b243-4c25-9aaf-c459ca926090 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.320551] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ee6acb7-043d-436d-8caf-f1aeab15b6d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.322951] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145874, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.155625} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.325734] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.326177] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 752.326177] env[66641]: value = "task-5145877" [ 752.326177] env[66641]: _type = "Task" [ 752.326177] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.330311] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83755359-ac98-48dd-9cd6-23a7ea021388 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.337360] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 752.337360] env[66641]: value = "task-5145878" [ 752.337360] env[66641]: _type = "Task" [ 752.337360] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.356627] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145877, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.370007] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Reconfiguring VM instance instance-0000001f to attach disk [datastore2] 0f6055b9-f5b4-48ba-9589-0af212808be7/0f6055b9-f5b4-48ba-9589-0af212808be7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.371035] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90a22d07-b7bc-4ba6-952d-1e81e2f0668b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.398480] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145878, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.410481] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145875, 'name': CreateVM_Task, 'duration_secs': 0.415037} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.412290] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.412762] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 752.412762] env[66641]: value = "task-5145879" [ 752.412762] env[66641]: _type = "Task" [ 752.412762] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.413864] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.414612] env[66641]: WARNING openstack [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.423276] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.423276] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.423276] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 752.424275] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-337b14b6-ceb2-450d-aa59-dcbe243657f7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.439065] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145879, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.441388] env[66641]: DEBUG nova.scheduler.client.report [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 752.445609] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 752.445609] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b64890-6f66-e0a7-e60b-44d2cea6baa2" [ 752.445609] env[66641]: _type = "Task" [ 752.445609] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.457611] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b64890-6f66-e0a7-e60b-44d2cea6baa2, 'name': SearchDatastore_Task, 'duration_secs': 0.017437} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.457966] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.458249] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.458504] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.458699] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.458901] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.459400] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1cc8ffd3-365d-4e3a-9e8c-50b14f74e259 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.468976] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.469248] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.470245] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-363fcd21-3e9d-4114-9d26-4aff072d978f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.481621] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 752.481621] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52647c43-be45-0467-1e71-3c8e5a753a75" [ 752.481621] env[66641]: _type = "Task" [ 752.481621] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.494223] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.494425] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 752.494705] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 752.496239] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52647c43-be45-0467-1e71-3c8e5a753a75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.529690] env[66641]: DEBUG oslo_concurrency.lockutils [req-dbb1649b-964d-468e-8892-0235efaa4b3b req-bb739bb0-d7c0-4098-a85f-a03dfe21a22e service nova] Releasing lock "refresh_cache-b9d032da-031e-42e0-86e2-95254c1ceac1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 752.537525] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.537932] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.746450] env[66641]: DEBUG nova.network.neutron [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Updating instance_info_cache with network_info: [{"id": "20765973-11e5-4b88-aa88-4c5760fa5d37", "address": "fa:16:3e:d4:24:46", "network": {"id": "d45e1e15-6956-41e2-a11d-acef7bf59f86", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1137265955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "400b5d331a5c48a0947adc3477fe65bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20765973-11", "ovs_interfaceid": "20765973-11e5-4b88-aa88-4c5760fa5d37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 752.854437] env[66641]: DEBUG oslo_vmware.api [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145878, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161897} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.857900] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 752.857900] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 752.858061] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 752.858561] env[66641]: INFO nova.compute.manager [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Took 1.19 seconds to destroy the instance on the hypervisor. [ 752.858561] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 752.858732] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145877, 'name': Rename_Task, 'duration_secs': 0.305267} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.859056] env[66641]: DEBUG nova.compute.manager [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 752.859095] env[66641]: DEBUG nova.network.neutron [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 752.859647] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.859896] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 752.868562] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 752.868562] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2ea863a1-dd19-43ce-9128-dff41c205e78 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.875443] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 752.875443] env[66641]: value = "task-5145880" [ 752.875443] env[66641]: _type = "Task" [ 752.875443] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.888529] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145880, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.903940] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.904239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.904477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 752.904669] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.904834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.907388] env[66641]: INFO nova.compute.manager [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Terminating instance [ 752.936088] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145879, 'name': ReconfigVM_Task, 'duration_secs': 0.337503} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.936397] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Reconfigured VM instance instance-0000001f to attach disk [datastore2] 0f6055b9-f5b4-48ba-9589-0af212808be7/0f6055b9-f5b4-48ba-9589-0af212808be7.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.937144] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ded3230-b145-4cda-b543-18aa20d35f4d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.945619] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 752.945619] env[66641]: value = "task-5145881" [ 752.945619] env[66641]: _type = "Task" [ 752.945619] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.946801] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 752.947045] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 752.954142] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.040s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 752.956104] env[66641]: INFO nova.compute.claims [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.968738] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145881, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.994064] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52647c43-be45-0467-1e71-3c8e5a753a75, 'name': SearchDatastore_Task, 'duration_secs': 0.016545} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.995061] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fce469a-61e6-42a2-a042-6aac764bea11 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.998259] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 752.998616] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.003762] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 753.010419] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 753.010419] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5290bf34-dc4d-a061-e4df-b89b488a151a" [ 753.010419] env[66641]: _type = "Task" [ 753.010419] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.024053] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5290bf34-dc4d-a061-e4df-b89b488a151a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.249822] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Releasing lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.251409] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Instance network_info: |[{"id": "20765973-11e5-4b88-aa88-4c5760fa5d37", "address": "fa:16:3e:d4:24:46", "network": {"id": "d45e1e15-6956-41e2-a11d-acef7bf59f86", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1137265955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "400b5d331a5c48a0947adc3477fe65bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20765973-11", "ovs_interfaceid": "20765973-11e5-4b88-aa88-4c5760fa5d37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 753.251607] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d4:24:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd88b750a-0e7d-4f16-8bd5-8e6d5743b720', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '20765973-11e5-4b88-aa88-4c5760fa5d37', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.259857] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Creating folder: Project (400b5d331a5c48a0947adc3477fe65bb). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 753.260314] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e8585d08-32e0-4b47-849c-efe83c1984df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.275022] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.275022] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.284542] env[66641]: DEBUG nova.network.neutron [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.293083] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Created folder: Project (400b5d331a5c48a0947adc3477fe65bb) in parent group-v1000566. [ 753.293083] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Creating folder: Instances. Parent ref: group-v1000656. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 753.293259] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5d105b12-6b15-4c9d-8be4-41afb2f1af0e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.299679] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.299960] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.311955] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Created folder: Instances in parent group-v1000656. [ 753.312634] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 753.312634] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 753.312770] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a7bc6b3-57ed-41a7-a52b-bb77c50aa2ce {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.340156] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.340156] env[66641]: value = "task-5145884" [ 753.340156] env[66641]: _type = "Task" [ 753.340156] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.359678] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145884, 'name': CreateVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.390795] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145880, 'name': PowerOnVM_Task} progress is 96%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.397388] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.398240] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.411498] env[66641]: DEBUG nova.compute.manager [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 753.411914] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.412926] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2676a4e-2dfe-4bbb-b9d6-f1568f080bf9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.423174] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 753.423594] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0059107-d698-4278-b724-aa5a3711f598 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.434265] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 753.434265] env[66641]: value = "task-5145885" [ 753.434265] env[66641]: _type = "Task" [ 753.434265] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.446441] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145885, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.449845] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Updated VIF entry in instance network info cache for port 298db8ec-67aa-4b6a-8981-4f1a72602ef3. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 753.449845] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Updating instance_info_cache with network_info: [{"id": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "address": "fa:16:3e:4f:a9:07", "network": {"id": "3adcc8d1-479f-4ffb-8202-b0f98664d188", "bridge": "br-int", "label": "tempest-ImagesOneServerNegativeTestJSON-416391934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "417e1e40aca447229001fa725c82e9ca", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4b5c60ce-845e-4506-bc10-348461fece6d", "external-id": "nsx-vlan-transportzone-831", "segmentation_id": 831, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap298db8ec-67", "ovs_interfaceid": "298db8ec-67aa-4b6a-8981-4f1a72602ef3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.456237] env[66641]: DEBUG nova.compute.utils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 753.458024] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 753.458144] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 753.458448] env[66641]: WARNING neutronclient.v2_0.client [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.458789] env[66641]: WARNING neutronclient.v2_0.client [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 753.459384] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.459737] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.475660] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145881, 'name': Rename_Task, 'duration_secs': 0.162927} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.480468] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 753.481771] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b761ee0-70ca-4ae4-a874-56f738d45bc2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.492040] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 753.492040] env[66641]: value = "task-5145886" [ 753.492040] env[66641]: _type = "Task" [ 753.492040] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.506529] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.523397] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5290bf34-dc4d-a061-e4df-b89b488a151a, 'name': SearchDatastore_Task, 'duration_secs': 0.016626} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.523791] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.524262] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] e508b1e9-cbb0-408c-98e1-d54ceaec1f1b/e508b1e9-cbb0-408c-98e1-d54ceaec1f1b.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.524636] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-652b814f-579d-45d6-8295-8b186d1dee6b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.534511] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 753.534511] env[66641]: value = "task-5145887" [ 753.534511] env[66641]: _type = "Task" [ 753.534511] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.547382] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145887, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.607513] env[66641]: DEBUG nova.policy [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '04d4575ad3ba4c938a1d1cb696365d3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b740816edb13463cb24e1caac7dbd569', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 753.656134] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.657840] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.789362] env[66641]: DEBUG nova.network.neutron [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Updating instance_info_cache with network_info: [{"id": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "address": "fa:16:3e:83:8b:97", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap923d6e74-64", "ovs_interfaceid": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 753.789996] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Releasing lock "refresh_cache-d03f39a4-532b-439a-9055-19fc1e769fff" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.790448] env[66641]: DEBUG nova.compute.manager [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 753.790634] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.791680] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75eed8c2-e97a-47c3-be7b-a73fae25df5d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.804852] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 753.805227] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56255a68-8b89-417f-9ae9-bf5e661a74fc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.817281] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 753.817281] env[66641]: value = "task-5145888" [ 753.817281] env[66641]: _type = "Task" [ 753.817281] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.828735] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145888, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.856485] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145884, 'name': CreateVM_Task, 'duration_secs': 0.480655} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.856485] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 753.857063] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 753.858021] env[66641]: WARNING openstack [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 753.865137] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.865137] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.865239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 753.866059] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1c014d0-e31a-4537-936e-d4d71043ea67 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.875025] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 753.875025] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258021f-6217-8b7e-278c-ddb7c2ab1827" [ 753.875025] env[66641]: _type = "Task" [ 753.875025] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.887163] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258021f-6217-8b7e-278c-ddb7c2ab1827, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.891031] env[66641]: DEBUG oslo_vmware.api [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145880, 'name': PowerOnVM_Task, 'duration_secs': 0.571133} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.891337] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 753.891553] env[66641]: INFO nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Took 12.64 seconds to spawn the instance on the hypervisor. [ 753.892455] env[66641]: DEBUG nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 753.892582] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f118b458-4e08-490b-a574-3c8ddf6a134e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.921210] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Received event network-vif-plugged-20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 753.921372] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquiring lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.921949] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.922648] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.922648] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] No waiting events found dispatching network-vif-plugged-20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 753.922648] env[66641]: WARNING nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Received unexpected event network-vif-plugged-20765973-11e5-4b88-aa88-4c5760fa5d37 for instance with vm_state building and task_state spawning. [ 753.922788] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Received event network-changed-20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 753.922837] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Refreshing instance network info cache due to event network-changed-20765973-11e5-4b88-aa88-4c5760fa5d37. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 753.923036] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquiring lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.923173] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquired lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.923320] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Refreshing network info cache for port 20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 753.950122] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145885, 'name': PowerOffVM_Task, 'duration_secs': 0.293116} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.950396] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 753.950566] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 753.950852] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-84125d22-2531-40ac-822f-4e515cfb8a3d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.958459] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Releasing lock "refresh_cache-0f6055b9-f5b4-48ba-9589-0af212808be7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 753.961696] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Received event network-vif-plugged-ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 753.961696] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquiring lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 753.961696] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 753.961696] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 753.961696] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] No waiting events found dispatching network-vif-plugged-ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 753.962228] env[66641]: WARNING nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Received unexpected event network-vif-plugged-ee7db498-4256-4265-9b73-6b6b2afa057c for instance with vm_state building and task_state spawning. [ 753.962228] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Received event network-changed-ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 753.962228] env[66641]: DEBUG nova.compute.manager [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Refreshing instance network info cache due to event network-changed-ee7db498-4256-4265-9b73-6b6b2afa057c. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 753.962228] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquiring lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.962228] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Acquired lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 753.962364] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Refreshing network info cache for port ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 753.968447] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 754.006660] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145886, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.024749] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Successfully created port: f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 754.044225] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 754.047076] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 754.047076] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleting the datastore file [datastore2] d14528cb-f26e-4fe5-8bf9-34900571fe03 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.047076] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aebed321-d5d5-422c-bd12-a89fad1cb402 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.054794] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145887, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.061423] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for the task: (returnval){ [ 754.061423] env[66641]: value = "task-5145890" [ 754.061423] env[66641]: _type = "Task" [ 754.061423] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.073983] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145890, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.118076] env[66641]: DEBUG nova.network.neutron [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 754.291826] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.292255] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Instance network_info: |[{"id": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "address": "fa:16:3e:83:8b:97", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap923d6e74-64", "ovs_interfaceid": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 754.292732] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:8b:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '923d6e74-642d-491f-bf2f-c3768d2d6e85', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 754.302715] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.305719] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 754.306493] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba98bcad-45fb-4e35-add9-f51653ec7922 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.336970] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145888, 'name': PowerOffVM_Task, 'duration_secs': 0.135173} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.339043] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 754.339043] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 754.339276] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 754.339276] env[66641]: value = "task-5145891" [ 754.339276] env[66641]: _type = "Task" [ 754.339276] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.339428] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1ddc8c10-1c23-4763-83c2-27546e02c7d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.354403] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145891, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.371921] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45998cbb-220d-48f7-a6f2-d6e60a745a7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.376585] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 754.376808] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 754.377531] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Deleting the datastore file [datastore1] d03f39a4-532b-439a-9055-19fc1e769fff {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.382222] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf8f084a-8aa4-42b1-9c04-a562e0ce50a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.393392] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b089b6-de14-438d-ab0a-b09623145550 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.397437] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5258021f-6217-8b7e-278c-ddb7c2ab1827, 'name': SearchDatastore_Task, 'duration_secs': 0.032757} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.399352] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.399656] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.399936] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.400150] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 754.400294] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.400660] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for the task: (returnval){ [ 754.400660] env[66641]: value = "task-5145893" [ 754.400660] env[66641]: _type = "Task" [ 754.400660] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.401357] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b100b2c3-3e14-44a8-9bcf-df01dfe00760 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.409399] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.458954] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.458954] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.459564] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.470641] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.472081] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.472422] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.479881] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2464447f-8de8-4ea2-93e4-fbe83b628492 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.484119] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.484119] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 754.488060] env[66641]: INFO nova.compute.manager [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Took 22.87 seconds to build instance. [ 754.490029] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.495154] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f888cebe-e111-466a-b009-2b49dbd2f67d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.497563] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145893, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.497563] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.497816] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.504701] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.504701] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 754.512038] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 754.514580] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44737fc7-8a1b-4816-819d-39daa218f866 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.519961] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 754.519961] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cfafb6-5e78-0af3-2426-0d74920ed5d4" [ 754.519961] env[66641]: _type = "Task" [ 754.519961] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.527166] env[66641]: DEBUG oslo_vmware.api [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145886, 'name': PowerOnVM_Task, 'duration_secs': 0.936091} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.532197] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 754.532294] env[66641]: INFO nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Took 10.79 seconds to spawn the instance on the hypervisor. [ 754.532454] env[66641]: DEBUG nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 754.543712] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a827f084-014f-4d16-a1da-156a6b3c35ce {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.547805] env[66641]: DEBUG nova.compute.provider_tree [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 754.565365] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cfafb6-5e78-0af3-2426-0d74920ed5d4, 'name': SearchDatastore_Task, 'duration_secs': 0.023307} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.574808] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f66ce565-8c6b-4fca-8e7f-6fe683945df8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.579035] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145887, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557594} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.586509] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] e508b1e9-cbb0-408c-98e1-d54ceaec1f1b/e508b1e9-cbb0-408c-98e1-d54ceaec1f1b.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.586788] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.587959] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8a22e890-3483-47ee-8386-bb0797208f52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.595465] env[66641]: DEBUG oslo_vmware.api [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Task: {'id': task-5145890, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.357334} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.595768] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 754.595768] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b4de57-dc68-95f7-c1b1-fa9702a9cfe7" [ 754.595768] env[66641]: _type = "Task" [ 754.595768] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.596722] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.596953] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 754.597257] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.597453] env[66641]: INFO nova.compute.manager [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Took 1.19 seconds to destroy the instance on the hypervisor. [ 754.597785] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.598200] env[66641]: DEBUG nova.compute.manager [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 754.598271] env[66641]: DEBUG nova.network.neutron [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 754.600035] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.600035] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.614738] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 754.614738] env[66641]: value = "task-5145894" [ 754.614738] env[66641]: _type = "Task" [ 754.614738] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.624505] env[66641]: INFO nova.compute.manager [-] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Took 1.77 seconds to deallocate network for instance. [ 754.624901] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b4de57-dc68-95f7-c1b1-fa9702a9cfe7, 'name': SearchDatastore_Task, 'duration_secs': 0.018401} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.632518] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 754.632518] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 25c7bd59-ec24-4d30-840b-3c4549dbf669/25c7bd59-ec24-4d30-840b-3c4549dbf669.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.632518] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-059fa1a1-843b-4341-9025-56d16a2dd5d7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.638192] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145894, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.649695] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 754.649695] env[66641]: value = "task-5145895" [ 754.649695] env[66641]: _type = "Task" [ 754.649695] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.663018] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145895, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.664483] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.664888] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.716031] env[66641]: DEBUG nova.compute.manager [req-287374f6-616d-4268-8cce-38ebcd622eb1 req-422795a9-5326-4a6c-adcf-8371740a86dc service nova] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Received event network-vif-deleted-dbb7738c-a759-4460-8f36-9f673e12b204 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 754.727850] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.728374] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.746538] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.746799] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.853263] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145891, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.914180] env[66641]: DEBUG oslo_vmware.api [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Task: {'id': task-5145893, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264295} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.914397] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 754.914573] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 754.914759] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 754.914972] env[66641]: INFO nova.compute.manager [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Took 1.12 seconds to destroy the instance on the hypervisor. [ 754.915193] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 754.915490] env[66641]: DEBUG nova.compute.manager [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 754.915490] env[66641]: DEBUG nova.network.neutron [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 754.916037] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 754.916312] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 754.990429] env[66641]: DEBUG oslo_concurrency.lockutils [None req-494ba81f-305a-4b0e-99c1-d0606a0d0eda tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.382s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 754.995989] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 755.014512] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.039986] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 755.040282] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 755.040797] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 755.041058] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 755.041215] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 755.041358] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 755.041605] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 755.041787] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 755.041938] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 755.042155] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 755.042421] env[66641]: DEBUG nova.virt.hardware [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 755.043711] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e79af80-3609-4914-97ae-f045c22b0de3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.052519] env[66641]: DEBUG nova.scheduler.client.report [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 755.057631] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9beffbd2-9ddc-45e2-be56-efa3c2638aaf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.097851] env[66641]: INFO nova.compute.manager [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Took 20.34 seconds to build instance. [ 755.131048] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145894, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.122609} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.131048] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.131048] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c56f1e3d-2a03-4370-a091-f20950fa5fe7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.136647] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.160736] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Reconfiguring VM instance instance-00000020 to attach disk [datastore2] e508b1e9-cbb0-408c-98e1-d54ceaec1f1b/e508b1e9-cbb0-408c-98e1-d54ceaec1f1b.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.165398] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92719fca-61e6-4c2f-be20-1f15349cd58d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.193048] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145895, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.195797] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 755.195797] env[66641]: value = "task-5145896" [ 755.195797] env[66641]: _type = "Task" [ 755.195797] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.209294] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145896, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.215568] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 755.215568] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.354175] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145891, 'name': CreateVM_Task, 'duration_secs': 0.809143} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.354375] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 755.355237] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.355606] env[66641]: WARNING openstack [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.362068] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.362337] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.362830] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 755.363210] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f077bc8-3181-4dc8-b246-7a2310e7d255 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.373974] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 755.373974] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5250381e-6bca-0bb0-ee78-df56dc815df6" [ 755.373974] env[66641]: _type = "Task" [ 755.373974] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.385670] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5250381e-6bca-0bb0-ee78-df56dc815df6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.510419] env[66641]: DEBUG nova.network.neutron [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 755.568348] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.614s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.568930] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 755.572238] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 12.072s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 755.572441] env[66641]: DEBUG nova.objects.instance [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 755.602740] env[66641]: DEBUG oslo_concurrency.lockutils [None req-29cef1bf-b73e-4cc5-8e89-749002c18fdd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.849s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 755.664649] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145895, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.914225} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.665016] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 25c7bd59-ec24-4d30-840b-3c4549dbf669/25c7bd59-ec24-4d30-840b-3c4549dbf669.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 755.665309] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 755.665638] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4aafb978-4a2d-4d28-b832-b56ac24cee35 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.676754] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 755.676754] env[66641]: value = "task-5145897" [ 755.676754] env[66641]: _type = "Task" [ 755.676754] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.686916] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.689721] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Successfully updated port: f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 755.708414] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145896, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.721853] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 755.741783] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.742279] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.770441] env[66641]: DEBUG nova.network.neutron [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 755.770790] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.771180] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.804217] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 755.804217] env[66641]: WARNING openstack [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 755.888136] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5250381e-6bca-0bb0-ee78-df56dc815df6, 'name': SearchDatastore_Task, 'duration_secs': 0.057245} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.888635] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 755.888983] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 755.889261] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.889400] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 755.889609] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 755.889977] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c88dbb62-72c3-48ff-ac88-8cc1bdbbff33 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.900401] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 755.900604] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 755.901434] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef503a60-27b9-4b1f-98cf-66859798afe3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.907562] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 755.907562] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5260ff8c-18c7-ed12-e11a-7d8fb77986e5" [ 755.907562] env[66641]: _type = "Task" [ 755.907562] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.918801] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5260ff8c-18c7-ed12-e11a-7d8fb77986e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.965068] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Updated VIF entry in instance network info cache for port 20765973-11e5-4b88-aa88-4c5760fa5d37. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 755.965446] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Updating instance_info_cache with network_info: [{"id": "20765973-11e5-4b88-aa88-4c5760fa5d37", "address": "fa:16:3e:d4:24:46", "network": {"id": "d45e1e15-6956-41e2-a11d-acef7bf59f86", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1137265955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "400b5d331a5c48a0947adc3477fe65bb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d88b750a-0e7d-4f16-8bd5-8e6d5743b720", "external-id": "nsx-vlan-transportzone-715", "segmentation_id": 715, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap20765973-11", "ovs_interfaceid": "20765973-11e5-4b88-aa88-4c5760fa5d37", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.016092] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Updated VIF entry in instance network info cache for port ee7db498-4256-4265-9b73-6b6b2afa057c. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 756.016437] env[66641]: DEBUG nova.network.neutron [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Updating instance_info_cache with network_info: [{"id": "ee7db498-4256-4265-9b73-6b6b2afa057c", "address": "fa:16:3e:f9:cb:a3", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.185", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapee7db498-42", "ovs_interfaceid": "ee7db498-4256-4265-9b73-6b6b2afa057c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.017791] env[66641]: INFO nova.compute.manager [-] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Took 1.42 seconds to deallocate network for instance. [ 756.076024] env[66641]: DEBUG nova.compute.utils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 756.076393] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 756.187952] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080198} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.188255] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 756.189068] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4fa9d13-b458-475f-afa7-896945010206 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.192139] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.192311] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.192466] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 756.216975] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 25c7bd59-ec24-4d30-840b-3c4549dbf669/25c7bd59-ec24-4d30-840b-3c4549dbf669.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 756.221082] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e8ac38d-7283-40c8-bcbb-4f853c754c13 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.247155] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145896, 'name': ReconfigVM_Task, 'duration_secs': 0.611657} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.247268] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Reconfigured VM instance instance-00000020 to attach disk [datastore2] e508b1e9-cbb0-408c-98e1-d54ceaec1f1b/e508b1e9-cbb0-408c-98e1-d54ceaec1f1b.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 756.248749] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-423706aa-0aab-4270-80a1-2ba7680d3a61 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.250965] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 756.250965] env[66641]: value = "task-5145898" [ 756.250965] env[66641]: _type = "Task" [ 756.250965] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.256911] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 756.256911] env[66641]: value = "task-5145899" [ 756.256911] env[66641]: _type = "Task" [ 756.256911] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.261934] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.265618] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145898, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.272642] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145899, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.279235] env[66641]: DEBUG nova.network.neutron [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 756.418339] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5260ff8c-18c7-ed12-e11a-7d8fb77986e5, 'name': SearchDatastore_Task, 'duration_secs': 0.017589} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.419244] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66278d4d-e03b-4be2-a584-3a139231b4e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.425912] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 756.425912] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522abe00-09aa-fe8d-3083-76b673e01d80" [ 756.425912] env[66641]: _type = "Task" [ 756.425912] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.435672] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522abe00-09aa-fe8d-3083-76b673e01d80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.473026] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Releasing lock "refresh_cache-25c7bd59-ec24-4d30-840b-3c4549dbf669" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.473026] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Received event network-vif-plugged-923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 756.473026] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquiring lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.473026] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.473026] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.473297] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] No waiting events found dispatching network-vif-plugged-923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 756.473297] env[66641]: WARNING nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Received unexpected event network-vif-plugged-923d6e74-642d-491f-bf2f-c3768d2d6e85 for instance with vm_state building and task_state spawning. [ 756.473297] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Received event network-changed-923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 756.473297] env[66641]: DEBUG nova.compute.manager [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Refreshing instance network info cache due to event network-changed-923d6e74-642d-491f-bf2f-c3768d2d6e85. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 756.473297] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquiring lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.473479] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Acquired lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 756.473479] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Refreshing network info cache for port 923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 756.527147] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.527563] env[66641]: DEBUG oslo_concurrency.lockutils [req-08472f3f-b2d3-48d5-afd2-ddc35ba72d48 req-6fd43c77-9fed-4487-8bb8-ee52cd1205be service nova] Releasing lock "refresh_cache-e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.582507] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 756.586642] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a75b380b-aec8-487c-b4b2-a7b6e53acb4a tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.587876] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.562s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.590159] env[66641]: INFO nova.compute.claims [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.696330] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.696814] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 756.702516] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 756.762285] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145898, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.772117] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145899, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.781799] env[66641]: INFO nova.compute.manager [-] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Took 1.87 seconds to deallocate network for instance. [ 756.892106] env[66641]: DEBUG nova.compute.manager [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Received event network-vif-plugged-f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 756.892750] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Acquiring lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 756.893012] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 756.894960] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 756.895223] env[66641]: DEBUG nova.compute.manager [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] No waiting events found dispatching network-vif-plugged-f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 756.895431] env[66641]: WARNING nova.compute.manager [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Received unexpected event network-vif-plugged-f6c1b16b-1415-4eaf-9228-e8c95836606c for instance with vm_state building and task_state spawning. [ 756.895619] env[66641]: DEBUG nova.compute.manager [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Received event network-changed-f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 756.895801] env[66641]: DEBUG nova.compute.manager [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Refreshing instance network info cache due to event network-changed-f6c1b16b-1415-4eaf-9228-e8c95836606c. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 756.896014] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Acquiring lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.938070] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522abe00-09aa-fe8d-3083-76b673e01d80, 'name': SearchDatastore_Task, 'duration_secs': 0.034775} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.938401] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 756.938742] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 05882781-78be-4568-95f4-2fccc4cf4dfe/05882781-78be-4568-95f4-2fccc4cf4dfe.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 756.938948] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-42056b79-b6bf-4573-90bb-0908ee630c7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.946992] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 756.946992] env[66641]: value = "task-5145900" [ 756.946992] env[66641]: _type = "Task" [ 756.946992] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.956678] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.976900] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 756.977317] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.070503] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.071179] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.262643] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145898, 'name': ReconfigVM_Task, 'duration_secs': 0.813662} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.266171] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 25c7bd59-ec24-4d30-840b-3c4549dbf669/25c7bd59-ec24-4d30-840b-3c4549dbf669.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 757.266858] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-51d6d98d-dc2e-48dc-b170-7433b05afbd3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.279872] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145899, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.281989] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 757.281989] env[66641]: value = "task-5145901" [ 757.281989] env[66641]: _type = "Task" [ 757.281989] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.288384] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 757.292402] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145901, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.324926] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.325315] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.347026] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.347586] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.356793] env[66641]: DEBUG nova.compute.manager [req-ee3af47a-640d-41ad-a7ce-6a87ae27731b req-3344a0ec-8e2c-492a-8c1d-a6b4233bf015 service nova] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Received event network-vif-deleted-f3fb5c31-b49d-4a8e-a045-cc8ce97c41e3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 757.462772] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.548889] env[66641]: DEBUG nova.network.neutron [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Updating instance_info_cache with network_info: [{"id": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "address": "fa:16:3e:4f:51:c1", "network": {"id": "3927588d-76db-4d5a-a950-9a3f2a771c1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-911103075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b740816edb13463cb24e1caac7dbd569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6c1b16b-14", "ovs_interfaceid": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.594661] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 757.595152] env[66641]: WARNING openstack [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 757.605783] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 757.653543] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 757.653872] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 757.659021] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 757.659021] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 757.659021] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 757.659021] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 757.659021] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 757.659472] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 757.659798] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 757.660065] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 757.660281] env[66641]: DEBUG nova.virt.hardware [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 757.664219] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f01780f-36c8-4d00-b757-74d4f6630a0d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.676299] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c63861-db0f-4d11-94c1-a2ba6dde641e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.700449] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 757.707799] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Creating folder: Project (9c7f8a6e5b2c4c9593f4d2eaf86eb5ac). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.717504] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-04c1ca5e-7124-468d-90b8-947648c30938 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.734448] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Created folder: Project (9c7f8a6e5b2c4c9593f4d2eaf86eb5ac) in parent group-v1000566. [ 757.736205] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Creating folder: Instances. Parent ref: group-v1000660. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 757.736205] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2150226c-e018-42c5-a418-2b3cee171c55 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.749556] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Created folder: Instances in parent group-v1000660. [ 757.749886] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 757.750204] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 757.750437] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-773cdb85-329d-4045-b775-692f5bc82d45 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.783094] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145899, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.783094] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 757.783094] env[66641]: value = "task-5145904" [ 757.783094] env[66641]: _type = "Task" [ 757.783094] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.802965] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145901, 'name': Rename_Task, 'duration_secs': 0.276555} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.803213] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Updated VIF entry in instance network info cache for port 923d6e74-642d-491f-bf2f-c3768d2d6e85. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 757.803561] env[66641]: DEBUG nova.network.neutron [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Updating instance_info_cache with network_info: [{"id": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "address": "fa:16:3e:83:8b:97", "network": {"id": "0d29fe16-a826-47e3-947f-701b29f83066", "bridge": "br-int", "label": "tempest-ImagesTestJSON-367786594-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17792883c6504d0f9f91fb6beb25b087", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b6942d7-d4ab-4b2a-8d0f-76bf2a2478ad", "external-id": "nsx-vlan-transportzone-871", "segmentation_id": 871, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap923d6e74-64", "ovs_interfaceid": "923d6e74-642d-491f-bf2f-c3768d2d6e85", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 757.804875] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 757.805364] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5d9238f-f459-450c-8f49-e870342c0f39 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.812149] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 757.812149] env[66641]: value = "task-5145905" [ 757.812149] env[66641]: _type = "Task" [ 757.812149] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.828358] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145905, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.967705] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145900, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.053301] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Releasing lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.053819] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Instance network_info: |[{"id": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "address": "fa:16:3e:4f:51:c1", "network": {"id": "3927588d-76db-4d5a-a950-9a3f2a771c1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-911103075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b740816edb13463cb24e1caac7dbd569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6c1b16b-14", "ovs_interfaceid": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 758.054290] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Acquired lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.054506] env[66641]: DEBUG nova.network.neutron [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Refreshing network info cache for port f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 758.055951] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:51:c1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ee617cec-01ea-4a11-ac04-ef9767f4c86d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f6c1b16b-1415-4eaf-9228-e8c95836606c', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 758.064844] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Creating folder: Project (b740816edb13463cb24e1caac7dbd569). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.066856] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.067128] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.074237] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-93c143cb-c75c-4db2-ae04-aefb55498014 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.093848] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Created folder: Project (b740816edb13463cb24e1caac7dbd569) in parent group-v1000566. [ 758.094206] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Creating folder: Instances. Parent ref: group-v1000663. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 758.095129] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c3f194f-d502-4d59-bdc4-746b7d7e4737 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.113858] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Created folder: Instances in parent group-v1000663. [ 758.114282] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 758.115543] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 758.116328] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d37d82-60de-4f50-a78c-dd2ecbf5b694 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.119739] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fc78ce6d-dcc8-4e75-bd77-e63478afc682 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.143838] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6eedb8-91cf-4f32-b56b-92006b9f15b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.151070] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 758.151070] env[66641]: value = "task-5145908" [ 758.151070] env[66641]: _type = "Task" [ 758.151070] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.194334] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 758.194617] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 758.196826] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145908, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.198100] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d065f1b6-89fb-44bf-a11c-e8d6ace6815d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.206854] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d82216ed-578c-49fc-b2fe-7ddb2d8a7ab0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.226541] env[66641]: DEBUG nova.compute.provider_tree [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.278963] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145899, 'name': Rename_Task, 'duration_secs': 1.700157} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.279468] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 758.279786] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51b6ca73-b905-4439-a6b0-023f08625f1b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.292844] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 758.292844] env[66641]: value = "task-5145909" [ 758.292844] env[66641]: _type = "Task" [ 758.292844] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.300491] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145904, 'name': CreateVM_Task, 'duration_secs': 0.470571} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.301137] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.301711] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.301897] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.302347] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 758.302705] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d9cf630-9d36-4d6b-bcc2-2ec9f0958d40 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.310057] env[66641]: DEBUG oslo_concurrency.lockutils [req-8ed7f78c-4fae-41ff-9292-329043b96511 req-ba6e3e34-278c-43a4-8773-eb08de1d69a6 service nova] Releasing lock "refresh_cache-05882781-78be-4568-95f4-2fccc4cf4dfe" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.310934] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145909, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.316991] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 758.316991] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5209b770-126b-a8fd-34cc-2634b9a0a314" [ 758.316991] env[66641]: _type = "Task" [ 758.316991] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.339756] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145905, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.344489] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5209b770-126b-a8fd-34cc-2634b9a0a314, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.389047] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.389047] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.470946] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145900, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.480945] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.481338] env[66641]: WARNING openstack [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.589196] env[66641]: DEBUG nova.network.neutron [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Updated VIF entry in instance network info cache for port f6c1b16b-1415-4eaf-9228-e8c95836606c. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 758.589602] env[66641]: DEBUG nova.network.neutron [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Updating instance_info_cache with network_info: [{"id": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "address": "fa:16:3e:4f:51:c1", "network": {"id": "3927588d-76db-4d5a-a950-9a3f2a771c1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-911103075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b740816edb13463cb24e1caac7dbd569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6c1b16b-14", "ovs_interfaceid": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 758.667214] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145908, 'name': CreateVM_Task, 'duration_secs': 0.50581} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.667433] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 758.671099] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 758.671099] env[66641]: WARNING openstack [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 758.677965] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.702513] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 758.735967] env[66641]: DEBUG nova.scheduler.client.report [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 758.807449] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145909, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.828403] env[66641]: DEBUG oslo_vmware.api [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145905, 'name': PowerOnVM_Task, 'duration_secs': 0.645693} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.829299] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 758.829426] env[66641]: INFO nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Took 10.72 seconds to spawn the instance on the hypervisor. [ 758.829590] env[66641]: DEBUG nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 758.830396] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3eaf98-c8ee-415a-9e32-534419ced535 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.836990] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5209b770-126b-a8fd-34cc-2634b9a0a314, 'name': SearchDatastore_Task, 'duration_secs': 0.094592} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.837595] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 758.837844] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 758.838104] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.838245] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.838419] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 758.838701] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 758.839020] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 758.839259] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-03cb0630-6d5c-4748-92b2-02cdc5e4aa86 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.842624] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-53adb70f-510f-40a9-9c6c-19139dd5f6b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.858350] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 758.858350] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa9fb4-1677-7b51-2cf8-6fec63d3844c" [ 758.858350] env[66641]: _type = "Task" [ 758.858350] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.864419] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 758.866080] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 758.866080] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0a6f8d8-dd76-48b8-bb94-e4e132b1e1ce {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.872295] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa9fb4-1677-7b51-2cf8-6fec63d3844c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.876846] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 758.876846] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520cb1bb-215c-4092-9300-3bada4da185f" [ 758.876846] env[66641]: _type = "Task" [ 758.876846] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.887124] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520cb1bb-215c-4092-9300-3bada4da185f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.970156] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145900, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.56791} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.970442] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 05882781-78be-4568-95f4-2fccc4cf4dfe/05882781-78be-4568-95f4-2fccc4cf4dfe.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 758.970650] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 758.971898] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0cc554cb-3b33-442e-8b4e-1a60512f835c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.979316] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 758.979316] env[66641]: value = "task-5145910" [ 758.979316] env[66641]: _type = "Task" [ 758.979316] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.990470] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145910, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.093771] env[66641]: DEBUG oslo_concurrency.lockutils [req-0c89f496-21cd-4983-aa7f-1c43addc9c0e req-221e71ce-d499-4b7f-875f-38172d68a13a service nova] Releasing lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.225375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.241950] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.654s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.242859] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 759.248365] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 7.956s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.310870] env[66641]: DEBUG oslo_vmware.api [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145909, 'name': PowerOnVM_Task, 'duration_secs': 0.851137} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.310870] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 759.310870] env[66641]: INFO nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Took 13.62 seconds to spawn the instance on the hypervisor. [ 759.310870] env[66641]: DEBUG nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 759.310870] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de56902a-1794-414c-ba7d-a735c0673b0a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.353579] env[66641]: DEBUG nova.compute.manager [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 759.361038] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a224f408-6e93-4b5e-b829-f7591b2fab9b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.367582] env[66641]: INFO nova.compute.manager [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Took 23.67 seconds to build instance. [ 759.378319] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aa9fb4-1677-7b51-2cf8-6fec63d3844c, 'name': SearchDatastore_Task, 'duration_secs': 0.016022} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.381681] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.382045] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 759.382140] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.389261] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520cb1bb-215c-4092-9300-3bada4da185f, 'name': SearchDatastore_Task, 'duration_secs': 0.017748} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.390036] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac63fbdd-25ba-4dfa-a605-3fe42b1fc9fb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.396079] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 759.396079] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52766176-dabf-4f5d-2905-2dbd49ca8751" [ 759.396079] env[66641]: _type = "Task" [ 759.396079] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.405844] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52766176-dabf-4f5d-2905-2dbd49ca8751, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.491312] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145910, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089045} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.491585] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 759.492408] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1d5b74-3fea-45fd-aa64-7bab8afa18d2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.523152] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Reconfiguring VM instance instance-00000022 to attach disk [datastore1] 05882781-78be-4568-95f4-2fccc4cf4dfe/05882781-78be-4568-95f4-2fccc4cf4dfe.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 759.523299] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b6c17cf-1c1b-4fa4-8236-0ab233fd841c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.545323] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 759.545323] env[66641]: value = "task-5145911" [ 759.545323] env[66641]: _type = "Task" [ 759.545323] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.554228] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145911, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.567016] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "8235441c-c046-49da-a5a5-92bd85e17982" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.567338] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.567570] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "8235441c-c046-49da-a5a5-92bd85e17982-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 759.567762] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 759.567978] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.570665] env[66641]: INFO nova.compute.manager [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Terminating instance [ 759.752915] env[66641]: DEBUG nova.compute.utils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 759.754360] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 759.757780] env[66641]: INFO nova.compute.claims [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.830251] env[66641]: INFO nova.compute.manager [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Took 24.96 seconds to build instance. [ 759.869721] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9b4184f5-634c-404b-a301-1190b2f7f097 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.188s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 759.880199] env[66641]: INFO nova.compute.manager [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] instance snapshotting [ 759.884444] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da82cbb7-eb0f-41c7-afe8-57f21a0611bf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.909201] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ea2968-7143-48f1-a46f-9e7a15b69850 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.924873] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52766176-dabf-4f5d-2905-2dbd49ca8751, 'name': SearchDatastore_Task, 'duration_secs': 0.0327} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.925647] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 759.925820] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 759.926835] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 759.926835] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 759.926835] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-123709d6-ee23-4221-b964-391c452739aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.929354] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-21c415ee-8b36-48b7-8a7d-cae7118c57e7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.937400] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 759.937400] env[66641]: value = "task-5145912" [ 759.937400] env[66641]: _type = "Task" [ 759.937400] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.943198] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 759.943963] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 759.947820] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a79e039-bb7f-4cd4-b6e7-c3c911482235 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.951517] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145912, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.956546] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 759.956546] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b36416-b421-fc6b-8841-f8134f7e12a0" [ 759.956546] env[66641]: _type = "Task" [ 759.956546] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.967708] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b36416-b421-fc6b-8841-f8134f7e12a0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.057731] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145911, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.078040] env[66641]: DEBUG nova.compute.manager [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 760.078566] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.082060] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e45e85f-59f3-4f94-8e82-6b175786cd85 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.094642] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 760.095391] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdd734a7-56a5-4aa8-89b2-646ec083e402 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.103688] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 760.103688] env[66641]: value = "task-5145913" [ 760.103688] env[66641]: _type = "Task" [ 760.103688] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.115205] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.265376] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 760.268094] env[66641]: INFO nova.compute.resource_tracker [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating resource usage from migration 553a2d88-0488-434e-be83-aa85f338f673 [ 760.335032] env[66641]: DEBUG oslo_concurrency.lockutils [None req-22800221-c0dc-4938-a511-84bfd7d12804 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.472s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 760.427886] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 760.428303] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d65501a0-37da-434f-8d87-529081fc8720 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.438999] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 760.438999] env[66641]: value = "task-5145914" [ 760.438999] env[66641]: _type = "Task" [ 760.438999] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.456728] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145912, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.462417] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145914, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.478033] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b36416-b421-fc6b-8841-f8134f7e12a0, 'name': SearchDatastore_Task, 'duration_secs': 0.019978} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.478519] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05f98e70-d6d9-4991-afc9-89b0e01f8004 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.486735] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 760.486735] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c0d7a0-7952-1445-c28f-d1f592d12bae" [ 760.486735] env[66641]: _type = "Task" [ 760.486735] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.501918] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c0d7a0-7952-1445-c28f-d1f592d12bae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.564395] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145911, 'name': ReconfigVM_Task, 'duration_secs': 0.565786} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.564565] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Reconfigured VM instance instance-00000022 to attach disk [datastore1] 05882781-78be-4568-95f4-2fccc4cf4dfe/05882781-78be-4568-95f4-2fccc4cf4dfe.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 760.565449] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9a0bdca8-dd6a-4ad5-b9e2-3f44d5226af3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.578915] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 760.578915] env[66641]: value = "task-5145915" [ 760.578915] env[66641]: _type = "Task" [ 760.578915] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.588900] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145915, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.601546] env[66641]: DEBUG nova.compute.manager [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Stashing vm_state: active {{(pid=66641) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 760.618201] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145913, 'name': PowerOffVM_Task, 'duration_secs': 0.316267} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.619293] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 760.619293] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 760.620028] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7c1ff2a3-28f3-4f3e-b068-e46b07da96a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.719610] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 760.719887] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 760.721093] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Deleting the datastore file [datastore2] 8235441c-c046-49da-a5a5-92bd85e17982 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.721430] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50a89021-e76a-4cac-b615-9233c373bd99 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.733150] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for the task: (returnval){ [ 760.733150] env[66641]: value = "task-5145917" [ 760.733150] env[66641]: _type = "Task" [ 760.733150] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.746717] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.785694] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1337eb86-7432-49a8-a0d6-33b2ccbbf319 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.797648] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5ad0fc-6fa4-4551-bded-9e1543dfa31e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.838451] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba7086d-2e00-40bd-a30b-25b76897fd9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.848288] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4ecf68-da8b-4bbe-96ad-c2b48d60d44c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.866886] env[66641]: DEBUG nova.compute.provider_tree [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 760.956150] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145912, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.783038} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.960509] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 760.960873] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 760.961214] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145914, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.961493] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-404c8097-39d2-46db-ae35-2b98362f49ab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.970482] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 760.970482] env[66641]: value = "task-5145918" [ 760.970482] env[66641]: _type = "Task" [ 760.970482] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.980617] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145918, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.000933] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c0d7a0-7952-1445-c28f-d1f592d12bae, 'name': SearchDatastore_Task, 'duration_secs': 0.067304} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.001296] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 761.001569] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/dad5ed23-71a1-4b55-856f-2484f8e62708.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 761.002624] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-12b205bc-f3ea-4eb8-8a52-882fd9bf3889 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.013694] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 761.013694] env[66641]: value = "task-5145919" [ 761.013694] env[66641]: _type = "Task" [ 761.013694] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.028768] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145919, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.103488] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145915, 'name': Rename_Task, 'duration_secs': 0.209764} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.103488] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 761.103488] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53490bd7-40bc-4a0e-888e-eb192bb42ef2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.117147] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 761.117147] env[66641]: value = "task-5145920" [ 761.117147] env[66641]: _type = "Task" [ 761.117147] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.129038] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145920, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.141633] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 761.247909] env[66641]: DEBUG oslo_vmware.api [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Task: {'id': task-5145917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266978} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.249264] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 761.250203] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 761.250470] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 761.250696] env[66641]: INFO nova.compute.manager [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Took 1.17 seconds to destroy the instance on the hypervisor. [ 761.252064] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 761.252064] env[66641]: DEBUG nova.compute.manager [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 761.252064] env[66641]: DEBUG nova.network.neutron [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 761.253120] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.254430] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.278841] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 761.314198] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 761.314576] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 761.314782] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 761.315035] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 761.315243] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 761.315427] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 761.315728] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.315891] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 761.316130] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 761.316362] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 761.316604] env[66641]: DEBUG nova.virt.hardware [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 761.317615] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6031c0b8-b1e7-4fd7-8828-ace42dad0ac6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.323706] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 761.324058] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 761.342548] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf783812-b5cf-43d5-8ab6-94e34d33df44 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.376134] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.382485] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Creating folder: Project (076797f59df643b4b30fc26b60e2189b). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.383259] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-76b7d64e-d9a2-40f1-81fc-a20bd0e691ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.402197] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Created folder: Project (076797f59df643b4b30fc26b60e2189b) in parent group-v1000566. [ 761.402197] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Creating folder: Instances. Parent ref: group-v1000666. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.402197] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0cc0b10f-b2fd-44f0-b921-fb94991bd739 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.412284] env[66641]: ERROR nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [req-68a49af3-4d37-412c-8aaa-24eea5771550] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 750ffd2d-5e46-4240-a614-995f2be7c9cb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-68a49af3-4d37-412c-8aaa-24eea5771550"}]} [ 761.417800] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Created folder: Instances in parent group-v1000666. [ 761.418496] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 761.418933] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.419137] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d51d2dfb-4f96-4fa1-b7ce-ed6dbcf818e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.437148] env[66641]: DEBUG nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Refreshing inventories for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 761.440998] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.440998] env[66641]: value = "task-5145923" [ 761.440998] env[66641]: _type = "Task" [ 761.440998] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.457104] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145923, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.457720] env[66641]: DEBUG nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Updating ProviderTree inventory for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 761.458132] env[66641]: DEBUG nova.compute.provider_tree [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 761.467239] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145914, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.485405] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145918, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097327} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.487030] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 761.487582] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f924f81c-211e-4fbf-8373-cf518de30648 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.493401] env[66641]: DEBUG nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Refreshing aggregate associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, aggregates: None {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 761.527681] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Reconfiguring VM instance instance-00000024 to attach disk [datastore1] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.527681] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1870795-b764-4256-ad6c-ba23af90cf1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.557056] env[66641]: DEBUG nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Refreshing trait associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, traits: HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 761.570818] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145919, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.572749] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 761.572749] env[66641]: value = "task-5145924" [ 761.572749] env[66641]: _type = "Task" [ 761.572749] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.588252] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145924, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.636567] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145920, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.952922] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145923, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.964330] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145914, 'name': CreateSnapshot_Task, 'duration_secs': 1.221766} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.964827] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 761.965622] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-629ac9dc-909b-4811-bcab-ab2eb7195aa5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.012287] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-674a6dde-e51c-41c9-a41f-932ee51fbcee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.020656] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b6e50f-b7a8-4370-8cda-84c65a5835ab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.060027] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ead371-09e4-418d-8e57-14989d4739c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.066415] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145919, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.606962} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.067261] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/dad5ed23-71a1-4b55-856f-2484f8e62708.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 762.067474] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 762.067742] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5c94b3e1-a57f-4b66-b78a-8fb805ee531b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.074041] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306bf402-936e-4a2c-b480-5102f0a86c2f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.081129] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 762.081129] env[66641]: value = "task-5145925" [ 762.081129] env[66641]: _type = "Task" [ 762.081129] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.108220] env[66641]: DEBUG nova.compute.provider_tree [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.118763] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145924, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.127590] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145925, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.135900] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145920, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.206945] env[66641]: DEBUG nova.compute.manager [req-0d454923-7e68-4095-8d84-b67a89a16b38 req-46efc6ef-a150-4731-a8d8-fcb974984202 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Received event network-vif-deleted-43a23cc1-3c20-4707-9152-ad24a94338b5 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 762.206945] env[66641]: INFO nova.compute.manager [req-0d454923-7e68-4095-8d84-b67a89a16b38 req-46efc6ef-a150-4731-a8d8-fcb974984202 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Neutron deleted interface 43a23cc1-3c20-4707-9152-ad24a94338b5; detaching it from the instance and deleting it from the info cache [ 762.206945] env[66641]: DEBUG nova.network.neutron [req-0d454923-7e68-4095-8d84-b67a89a16b38 req-46efc6ef-a150-4731-a8d8-fcb974984202 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 762.281856] env[66641]: DEBUG nova.network.neutron [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 762.453765] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145923, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.489363] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 762.490378] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3edca6c5-64ed-4269-a6a1-857325f99e74 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.501060] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 762.501060] env[66641]: value = "task-5145926" [ 762.501060] env[66641]: _type = "Task" [ 762.501060] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.512374] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145926, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.591258] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145924, 'name': ReconfigVM_Task, 'duration_secs': 0.906202} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.595245] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Reconfigured VM instance instance-00000024 to attach disk [datastore1] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 762.596506] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff44de9b-95b7-43bf-91fb-a4fa5567598d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.607880] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145925, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146247} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.608315] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 762.610085] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d34daa7-3156-4093-8d73-ff1c53937ae1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.615471] env[66641]: DEBUG nova.scheduler.client.report [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 762.620423] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 762.620423] env[66641]: value = "task-5145927" [ 762.620423] env[66641]: _type = "Task" [ 762.620423] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.659444] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/dad5ed23-71a1-4b55-856f-2484f8e62708.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 762.661868] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad3daeb1-9cb1-493d-837e-2f9ef34919a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.695488] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145927, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.696118] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145920, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.704088] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 762.704088] env[66641]: value = "task-5145928" [ 762.704088] env[66641]: _type = "Task" [ 762.704088] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.710640] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-454b763b-01c1-4ca8-a766-3e6e4b6bef34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.718157] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145928, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.726682] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21abaaa8-5d86-428a-b3db-c79af016629c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.778025] env[66641]: DEBUG nova.compute.manager [req-0d454923-7e68-4095-8d84-b67a89a16b38 req-46efc6ef-a150-4731-a8d8-fcb974984202 service nova] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Detach interface failed, port_id=43a23cc1-3c20-4707-9152-ad24a94338b5, reason: Instance 8235441c-c046-49da-a5a5-92bd85e17982 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 762.785148] env[66641]: INFO nova.compute.manager [-] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Took 1.53 seconds to deallocate network for instance. [ 762.953995] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145923, 'name': CreateVM_Task, 'duration_secs': 1.071188} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.954419] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 762.954852] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.955052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 762.955390] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 762.955677] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2fceb159-c64b-4a1c-a1a0-ad88968b659c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.963096] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.963096] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.963219] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 762.963400] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 762.963546] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 762.965273] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 762.965273] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f1eadd-3fc4-76f4-cc99-7f7fe04374a6" [ 762.965273] env[66641]: _type = "Task" [ 762.965273] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.965916] env[66641]: INFO nova.compute.manager [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Terminating instance [ 762.978287] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f1eadd-3fc4-76f4-cc99-7f7fe04374a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.013588] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145926, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.122251] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 3.875s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.122654] env[66641]: INFO nova.compute.manager [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Migrating [ 763.131674] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.117s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.132094] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 763.132181] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 763.132430] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.996s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 763.132639] env[66641]: DEBUG nova.objects.instance [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lazy-loading 'resources' on Instance uuid fb1bf80c-b243-4c25-9aaf-c459ca926090 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.145698] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c61dc076-ca16-4ff4-8c30-a4dd693ce393 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.166399] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145927, 'name': Rename_Task, 'duration_secs': 0.319973} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.172556] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 763.173024] env[66641]: DEBUG oslo_vmware.api [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145920, 'name': PowerOnVM_Task, 'duration_secs': 1.683047} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.174148] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ff47685d-fc91-4ec3-b9ac-72b413e24b7c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.177053] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162b809a-6353-46e1-8172-f5a14af5645b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.183104] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 763.183305] env[66641]: INFO nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Took 13.20 seconds to spawn the instance on the hypervisor. [ 763.183481] env[66641]: DEBUG nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 763.185268] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd3115e-30e5-4b33-a64e-0af4ee171233 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.205163] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71772987-2d0e-47ec-b8a9-9c341b8c09b2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.213338] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 763.213338] env[66641]: value = "task-5145929" [ 763.213338] env[66641]: _type = "Task" [ 763.213338] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.226925] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293cd00d-d336-426f-8f35-f17bbacff7d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.240265] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145928, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.240265] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145929, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.271841] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179403MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 763.272303] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.292721] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 763.473829] env[66641]: DEBUG nova.compute.manager [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 763.474120] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 763.474902] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cd47f9-dc42-4bca-af9e-55a1101d548c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.487045] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f1eadd-3fc4-76f4-cc99-7f7fe04374a6, 'name': SearchDatastore_Task, 'duration_secs': 0.031799} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.489499] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 763.490346] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 763.490346] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.490522] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.490602] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 763.490916] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 763.491166] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cec75365-8ec6-4ad2-80a2-870456e37feb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.494535] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-2bf334f2-8ef7-4819-81d2-16cbbdc7c764 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.507467] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 763.507645] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 763.509408] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836f55bc-a693-4183-83dd-e95cd3557379 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.514465] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145926, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.518646] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 763.518646] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231e063-3978-5c9a-facf-1137633fd09b" [ 763.518646] env[66641]: _type = "Task" [ 763.518646] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.529592] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231e063-3978-5c9a-facf-1137633fd09b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.594498] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 763.594498] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 763.594498] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleting the datastore file [datastore2] 530f1e0e-5911-434e-bf20-edfd9778d7ad {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 763.594498] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-679001e2-34c4-49fd-b4e0-7f08d8df7f55 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.602595] env[66641]: DEBUG oslo_vmware.api [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 763.602595] env[66641]: value = "task-5145931" [ 763.602595] env[66641]: _type = "Task" [ 763.602595] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.612059] env[66641]: DEBUG oslo_vmware.api [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.661028] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.661028] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 763.661028] env[66641]: DEBUG nova.network.neutron [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 763.736163] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145928, 'name': ReconfigVM_Task, 'duration_secs': 0.669398} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.740306] env[66641]: INFO nova.compute.manager [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Took 26.92 seconds to build instance. [ 763.741516] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Reconfigured VM instance instance-00000023 to attach disk [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/dad5ed23-71a1-4b55-856f-2484f8e62708.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 763.747307] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e05882f7-2427-4a13-9176-1a26644b2437 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.749280] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145929, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.759485] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 763.759485] env[66641]: value = "task-5145932" [ 763.759485] env[66641]: _type = "Task" [ 763.759485] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.772674] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145932, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.014841] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145926, 'name': CloneVM_Task} progress is 95%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.030165] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231e063-3978-5c9a-facf-1137633fd09b, 'name': SearchDatastore_Task, 'duration_secs': 0.017028} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.030941] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-eeae6824-e382-4fd8-8188-0eeaf0724368 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.039455] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 764.039455] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52487147-6bea-d54a-9194-2c9744d6103b" [ 764.039455] env[66641]: _type = "Task" [ 764.039455] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.060128] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52487147-6bea-d54a-9194-2c9744d6103b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.116804] env[66641]: DEBUG oslo_vmware.api [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5145931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246109} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.116804] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 764.116804] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 764.116804] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 764.116804] env[66641]: INFO nova.compute.manager [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Took 0.64 seconds to destroy the instance on the hypervisor. [ 764.117118] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 764.117307] env[66641]: DEBUG nova.compute.manager [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 764.117428] env[66641]: DEBUG nova.network.neutron [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 764.118066] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.118364] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.166951] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.167378] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.177584] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.177877] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.197456] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10fa2dcc-e146-4ced-84ae-b60fdfb7dcdf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.205588] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d65c2c74-c7cf-4a98-bbb8-5a8273af81e9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.244893] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e6d3aac-9c2d-47c0-8240-66d5ae12b3dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.252993] env[66641]: DEBUG oslo_concurrency.lockutils [None req-a6a144da-7067-44ea-bd25-c481a411f9d1 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.452s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 764.262942] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef9e55e-5924-403b-9f54-97ff36954157 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.267130] env[66641]: DEBUG oslo_vmware.api [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145929, 'name': PowerOnVM_Task, 'duration_secs': 0.599248} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.270788] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 764.271033] env[66641]: INFO nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Took 6.67 seconds to spawn the instance on the hypervisor. [ 764.271635] env[66641]: DEBUG nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 764.276019] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471900e8-f9a0-4886-93aa-c2b945467702 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.285758] env[66641]: DEBUG nova.compute.manager [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 764.286717] env[66641]: DEBUG nova.compute.provider_tree [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.294217] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52f1e957-33aa-4b34-aff9-8661219fbaaa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.294718] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145932, 'name': Rename_Task, 'duration_secs': 0.211469} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.298563] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 764.303649] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a8bafe1e-5450-43e9-988d-967f7046976d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.313481] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 764.313481] env[66641]: value = "task-5145933" [ 764.313481] env[66641]: _type = "Task" [ 764.313481] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.322478] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145933, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.483684] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 764.484286] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 764.514666] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145926, 'name': CloneVM_Task, 'duration_secs': 1.790833} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.515058] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Created linked-clone VM from snapshot [ 764.515834] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb55fea-c231-4cb4-854d-d369a405e20b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.524504] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Uploading image 9fc435a6-c36e-48ab-88ef-6474a7187c18 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 764.535752] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 764.536150] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e3035b2a-56e1-4964-825a-5e7f2105eb21 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.544409] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 764.544409] env[66641]: value = "task-5145934" [ 764.544409] env[66641]: _type = "Task" [ 764.544409] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.557977] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52487147-6bea-d54a-9194-2c9744d6103b, 'name': SearchDatastore_Task, 'duration_secs': 0.017632} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.561418] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 764.561726] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 764.562043] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145934, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.562291] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0e8bfde8-4d06-4b9a-84e9-503e993b2b28 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.569948] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 764.569948] env[66641]: value = "task-5145935" [ 764.569948] env[66641]: _type = "Task" [ 764.569948] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.579856] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145935, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.800031] env[66641]: DEBUG nova.scheduler.client.report [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 764.816756] env[66641]: INFO nova.compute.manager [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] instance snapshotting [ 764.818733] env[66641]: INFO nova.compute.manager [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Took 21.92 seconds to build instance. [ 764.824861] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d28ac5c-e4e7-48c7-b41d-c2565cc4d9e3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.839310] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145933, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.857026] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e6cf9d-3edb-40a7-aef3-dacac3f82ffd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.968171] env[66641]: DEBUG nova.network.neutron [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 765.063834] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145934, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.089432] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145935, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.304281] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.172s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.309816] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.046s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 765.309816] env[66641]: INFO nova.compute.claims [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.324517] env[66641]: DEBUG oslo_concurrency.lockutils [None req-30f16a88-202f-4f67-95d9-89d8f37d858b tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.442s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.331208] env[66641]: DEBUG oslo_vmware.api [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145933, 'name': PowerOnVM_Task, 'duration_secs': 0.79228} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.331208] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 765.331500] env[66641]: INFO nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Took 10.34 seconds to spawn the instance on the hypervisor. [ 765.331564] env[66641]: DEBUG nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 765.332841] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d942edb2-5df5-437e-8ffd-8d361b1aa7b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.355807] env[66641]: INFO nova.scheduler.client.report [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted allocations for instance fb1bf80c-b243-4c25-9aaf-c459ca926090 [ 765.370336] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 765.371027] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c7b9e48d-617d-4f3c-a013-439a04dabe99 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.380770] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 765.380770] env[66641]: value = "task-5145936" [ 765.380770] env[66641]: _type = "Task" [ 765.380770] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.396699] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145936, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.433538] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 765.433538] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 765.471504] env[66641]: INFO nova.compute.manager [-] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Took 1.35 seconds to deallocate network for instance. [ 765.560150] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145934, 'name': Destroy_Task, 'duration_secs': 0.897223} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.560432] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Destroyed the VM [ 765.560665] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 765.561050] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c7214ab6-1d3c-48e2-ba96-35cd2ef6c6f4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.570281] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 765.570281] env[66641]: value = "task-5145937" [ 765.570281] env[66641]: _type = "Task" [ 765.570281] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.584684] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145937, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.587909] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145935, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690568} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.588230] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 765.588443] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 765.589128] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-711fb903-011c-40d6-8036-c2136291a6f6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.598143] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 765.598143] env[66641]: value = "task-5145938" [ 765.598143] env[66641]: _type = "Task" [ 765.598143] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.611110] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.863722] env[66641]: INFO nova.compute.manager [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Took 23.58 seconds to build instance. [ 765.868857] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c71b314b-d082-4c83-88e4-d7557d76c450 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "fb1bf80c-b243-4c25-9aaf-c459ca926090" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.710s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 765.894953] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145936, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.978798] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 766.014208] env[66641]: DEBUG nova.network.neutron [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 766.084431] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145937, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.111586] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100794} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.111953] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.112891] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3dac7fe-c14c-451d-b046-f9f6d091abf8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.139112] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.139403] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9bcb0a59-30a8-4ce4-a5dc-a020de3c591c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.161453] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 766.161453] env[66641]: value = "task-5145939" [ 766.161453] env[66641]: _type = "Task" [ 766.161453] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.173615] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145939, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.212790] env[66641]: DEBUG nova.compute.manager [req-2a8a74c5-c675-4a82-955b-ee2e3acc4992 req-c69ef64a-ccef-4b84-945f-6f34ccaa838f service nova] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Received event network-vif-deleted-6d596bf0-768d-4562-9bc7-6349b3d94101 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 766.370311] env[66641]: DEBUG oslo_concurrency.lockutils [None req-42115b29-3110-4ebd-bdaa-98439c87d367 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.101s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 766.396044] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145936, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.518451] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 766.589793] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145937, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.675353] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145939, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.765258] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48186a88-de3c-4793-999e-44f73b63eda0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.777879] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df1d6893-0307-483f-801b-ff0d64f8788e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.811110] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6989c338-e3ea-476b-9389-397737061428 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.821458] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f08f614-f0e1-43ab-86ee-ba418049da81 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.838833] env[66641]: DEBUG nova.compute.provider_tree [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 766.898278] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145936, 'name': CreateSnapshot_Task, 'duration_secs': 1.06768} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.898278] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 766.899075] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffabda77-7f58-4f90-af88-2f4e90c88f3f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.977344] env[66641]: DEBUG nova.compute.manager [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 766.978371] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9fbb2bf-62b9-402f-8f7a-cdf4dcdd27e8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.089274] env[66641]: DEBUG oslo_vmware.api [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145937, 'name': RemoveSnapshot_Task, 'duration_secs': 1.12961} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.089274] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 767.173603] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145939, 'name': ReconfigVM_Task, 'duration_secs': 0.569336} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.173937] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.174642] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bc2640df-4d3b-4235-ac17-6dae67d52088 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.183129] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 767.183129] env[66641]: value = "task-5145940" [ 767.183129] env[66641]: _type = "Task" [ 767.183129] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.196346] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145940, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.342592] env[66641]: DEBUG nova.scheduler.client.report [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 767.422547] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 767.423776] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b439cd6d-9636-41cd-93ef-a4eea21df7b2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.435720] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 767.435720] env[66641]: value = "task-5145941" [ 767.435720] env[66641]: _type = "Task" [ 767.435720] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.447838] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145941, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.467758] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "207483d3-803c-495b-9b93-6f986f3ca56e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.468019] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.491423] env[66641]: INFO nova.compute.manager [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] instance snapshotting [ 767.496550] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e04465b-5517-4c16-96e0-43eb3413a1eb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.523282] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4bbddf9-a32b-43a5-a917-0ce4ca2442a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.593714] env[66641]: WARNING nova.compute.manager [None req-593cdbff-3731-4c1f-b816-3451a9a8faf5 tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Image not found during snapshot: nova.exception.ImageNotFound: Image 9fc435a6-c36e-48ab-88ef-6474a7187c18 could not be found. [ 767.695624] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145940, 'name': Rename_Task, 'duration_secs': 0.380683} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.699357] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.701018] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f3f5964a-e772-45d6-96b6-cb47362762ff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.715300] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 767.715300] env[66641]: value = "task-5145942" [ 767.715300] env[66641]: _type = "Task" [ 767.715300] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.725431] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145942, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.727243] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.727706] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.728113] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 767.730106] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.730106] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.731357] env[66641]: INFO nova.compute.manager [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Terminating instance [ 767.849376] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 767.850235] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 767.854238] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.327s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 767.854660] env[66641]: DEBUG nova.objects.instance [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lazy-loading 'resources' on Instance uuid d14528cb-f26e-4fe5-8bf9-34900571fe03 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.908503] env[66641]: INFO nova.compute.manager [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Rescuing [ 767.908898] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.909024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 767.909211] env[66641]: DEBUG nova.network.neutron [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 767.947155] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145941, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.971124] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 768.049018] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 768.049890] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3856b139-d75f-4850-b0ef-20c0fdb93d5b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.053428] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-fb7ab37e-5331-4d5b-b771-e94b5118f589 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.072974] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 0 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 768.080684] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 768.080684] env[66641]: value = "task-5145943" [ 768.080684] env[66641]: _type = "Task" [ 768.080684] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.092382] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145943, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.228259] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145942, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.236830] env[66641]: DEBUG nova.compute.manager [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 768.236830] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 768.237922] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f854be7-b8f5-41ea-864d-9c1f09aba9c0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.249072] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.249496] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dedb52e6-faa5-4f01-adfd-51bcfe8d3b45 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.260060] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 768.260060] env[66641]: value = "task-5145944" [ 768.260060] env[66641]: _type = "Task" [ 768.260060] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.272228] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145944, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.362029] env[66641]: DEBUG nova.compute.utils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 768.364594] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 768.364594] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 768.364969] env[66641]: WARNING neutronclient.v2_0.client [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.366395] env[66641]: WARNING neutronclient.v2_0.client [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 768.366395] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.366395] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.412873] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 768.413487] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 768.450388] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145941, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.468676] env[66641]: DEBUG nova.policy [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ba7ff9aa8f34a7da6344773cea86836', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b2c8481015524aee95a933f61082faec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 768.501392] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 768.580108] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 768.580413] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-401e5b80-6de9-44a3-96ad-8f29c0f2d457 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.591814] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145943, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.593293] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 768.593293] env[66641]: value = "task-5145945" [ 768.593293] env[66641]: _type = "Task" [ 768.593293] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.606660] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.732568] env[66641]: DEBUG oslo_vmware.api [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145942, 'name': PowerOnVM_Task, 'duration_secs': 0.589472} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.732916] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 768.733199] env[66641]: INFO nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Took 7.45 seconds to spawn the instance on the hypervisor. [ 768.733460] env[66641]: DEBUG nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 768.734646] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ab9b93-51f0-4652-a52c-1ce1214a2ebf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.775954] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145944, 'name': PowerOffVM_Task, 'duration_secs': 0.265065} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.780784] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 768.781205] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 768.782151] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c4d987fb-b3c4-43d3-9993-dcf795ec0abd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.804623] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f050219c-e7c5-4bc9-ada2-b52c81ec62cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.816330] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4407357-ef8c-4dd8-91e8-1d6ab7c19554 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.870524] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50243e9f-9ffd-4594-b3f4-c0e83608b5b0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.876044] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 768.884944] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db0628c-dc5c-4830-bf1a-73f9fd731477 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.895036] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 768.895036] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 768.895036] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Deleting the datastore file [datastore1] 25c7bd59-ec24-4d30-840b-3c4549dbf669 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 768.895036] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d8797ccc-9eba-4f7c-8359-e777962725ac {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.912437] env[66641]: DEBUG nova.compute.provider_tree [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 768.916051] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for the task: (returnval){ [ 768.916051] env[66641]: value = "task-5145947" [ 768.916051] env[66641]: _type = "Task" [ 768.916051] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.928535] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.947626] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145941, 'name': CloneVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.044913] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Successfully created port: cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 769.096836] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145943, 'name': CreateSnapshot_Task, 'duration_secs': 0.816349} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.100781] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 769.101317] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0799da4-c33d-4151-a5ca-ad4c99b72790 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.118016] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145945, 'name': PowerOffVM_Task, 'duration_secs': 0.309049} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.121738] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 769.121738] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 17 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 769.129326] env[66641]: INFO nova.compute.manager [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Rebuilding instance [ 769.191238] env[66641]: DEBUG nova.compute.manager [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 769.192082] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4317ea36-0735-4666-8d57-92b1dd32352a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.270403] env[66641]: INFO nova.compute.manager [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Took 22.31 seconds to build instance. [ 769.380182] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.381255] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.418537] env[66641]: DEBUG nova.scheduler.client.report [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 769.432895] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.449689] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145941, 'name': CloneVM_Task, 'duration_secs': 1.566793} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.450547] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Created linked-clone VM from snapshot [ 769.450834] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57efc1c-10e6-4f2e-b746-bfce9283751c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.460617] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Uploading image a21499e1-a4d6-4fad-b53b-341f9044cc53 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 769.492190] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 769.492190] env[66641]: value = "vm-1000672" [ 769.492190] env[66641]: _type = "VirtualMachine" [ 769.492190] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 769.492479] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-67d08d67-0478-48d1-9f1b-6ed00d338d79 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.501130] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease: (returnval){ [ 769.501130] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528d6196-e61f-478f-87c6-4e15e5e4e011" [ 769.501130] env[66641]: _type = "HttpNfcLease" [ 769.501130] env[66641]: } obtained for exporting VM: (result){ [ 769.501130] env[66641]: value = "vm-1000672" [ 769.501130] env[66641]: _type = "VirtualMachine" [ 769.501130] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 769.501941] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the lease: (returnval){ [ 769.501941] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528d6196-e61f-478f-87c6-4e15e5e4e011" [ 769.501941] env[66641]: _type = "HttpNfcLease" [ 769.501941] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 769.508817] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 769.508817] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528d6196-e61f-478f-87c6-4e15e5e4e011" [ 769.508817] env[66641]: _type = "HttpNfcLease" [ 769.508817] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 769.626739] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 769.628636] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 769.628889] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 769.629773] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 769.629773] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 769.629773] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 769.629773] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 769.629773] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.629994] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 769.630193] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 769.630763] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 769.630763] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 769.636084] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-44ba7c2d-cf1c-44b5-b76d-d3f9a36b81ff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.639874] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9f0c9a31-5a7a-48de-96d2-bdaefb5163b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.659274] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 769.659274] env[66641]: value = "task-5145949" [ 769.659274] env[66641]: _type = "Task" [ 769.659274] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.661821] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 769.661821] env[66641]: value = "task-5145950" [ 769.661821] env[66641]: _type = "Task" [ 769.661821] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.675944] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145950, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.679167] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145949, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.773656] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ad614340-4acd-4579-8f90-89537f9a9806 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.818s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.896992] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 769.927471] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 769.927719] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 769.927866] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 769.928057] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 769.928209] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 769.928382] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 769.928609] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 769.928763] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 769.928924] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 769.929168] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 769.929354] env[66641]: DEBUG nova.virt.hardware [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 769.930177] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.076s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 769.932654] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19640d15-c925-456c-ae89-a21d02ae3f60 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.935905] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.648s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 769.936139] env[66641]: DEBUG nova.objects.instance [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lazy-loading 'resources' on Instance uuid d03f39a4-532b-439a-9055-19fc1e769fff {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 769.947192] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c0a94c4-4707-42a9-97c1-93ce88bd3204 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.951350] env[66641]: DEBUG oslo_vmware.api [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Task: {'id': task-5145947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.59012} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.952734] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 769.954254] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 769.954254] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 769.954254] env[66641]: INFO nova.compute.manager [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Took 1.72 seconds to destroy the instance on the hypervisor. [ 769.954254] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 769.954484] env[66641]: DEBUG nova.compute.manager [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 769.954589] env[66641]: DEBUG nova.network.neutron [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 769.955160] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 769.955443] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 769.962741] env[66641]: INFO nova.scheduler.client.report [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Deleted allocations for instance d14528cb-f26e-4fe5-8bf9-34900571fe03 [ 770.011108] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 770.011108] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528d6196-e61f-478f-87c6-4e15e5e4e011" [ 770.011108] env[66641]: _type = "HttpNfcLease" [ 770.011108] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 770.011476] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 770.011476] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528d6196-e61f-478f-87c6-4e15e5e4e011" [ 770.011476] env[66641]: _type = "HttpNfcLease" [ 770.011476] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 770.012389] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f245a09-8160-4967-be26-e8e722b38df2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.024922] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 770.025133] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 770.154685] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-782ac4df-2854-41e1-8c50-b76370c3f9b2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.173412] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145949, 'name': CloneVM_Task} progress is 93%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.181703] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145950, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.210025] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 770.210025] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64b121fb-7762-475a-9058-593f9e947db9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.219176] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 770.219176] env[66641]: value = "task-5145951" [ 770.219176] env[66641]: _type = "Task" [ 770.219176] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.231379] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145951, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.348265] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.348851] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.363342] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 770.367024] env[66641]: WARNING openstack [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 770.480257] env[66641]: DEBUG oslo_concurrency.lockutils [None req-db590542-a0ea-4dc7-8712-c00e99b83200 tempest-ListServersNegativeTestJSON-896808564 tempest-ListServersNegativeTestJSON-896808564-project-member] Lock "d14528cb-f26e-4fe5-8bf9-34900571fe03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 17.575s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 770.681248] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145949, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.687630] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145950, 'name': ReconfigVM_Task, 'duration_secs': 0.955798} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.691222] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 33 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 770.740317] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145951, 'name': PowerOffVM_Task, 'duration_secs': 0.211926} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.740317] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 770.741331] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.742654] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9faec59-a116-4a7e-b86f-c2d667f4f9a9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.754621] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 770.754992] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1d03bbe-e305-4c22-b2d9-ae2b4bb02e88 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.783523] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 770.786087] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 770.786087] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Deleting the datastore file [datastore1] 61042df0-a727-4aa8-b2ea-bdc40899d0fc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.786087] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2245a67-cd05-4c24-97bc-788a9e03dc91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.798734] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 770.798734] env[66641]: value = "task-5145953" [ 770.798734] env[66641]: _type = "Task" [ 770.798734] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.811681] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145953, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.846253] env[66641]: DEBUG nova.network.neutron [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Updating instance_info_cache with network_info: [{"id": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "address": "fa:16:3e:4f:51:c1", "network": {"id": "3927588d-76db-4d5a-a950-9a3f2a771c1c", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-911103075-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "b740816edb13463cb24e1caac7dbd569", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ee617cec-01ea-4a11-ac04-ef9767f4c86d", "external-id": "nsx-vlan-transportzone-11", "segmentation_id": 11, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf6c1b16b-14", "ovs_interfaceid": "f6c1b16b-1415-4eaf-9228-e8c95836606c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 770.869418] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad481257-bc29-40e2-9fe3-23ab50f1a25d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.880021] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c68b235-9b0e-4f42-b5b0-268aeabbc6c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.930702] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e172746c-586f-4dae-9e2e-1575bac6012a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.941962] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-406edff3-a665-42c1-b173-18db0e598414 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.960831] env[66641]: DEBUG nova.compute.provider_tree [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.997989] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Successfully updated port: cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 771.160344] env[66641]: DEBUG nova.network.neutron [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 771.178460] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145949, 'name': CloneVM_Task} progress is 95%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.201102] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 771.202067] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 771.202067] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 771.202558] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 771.202943] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 771.203600] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 771.203801] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.204702] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 771.204702] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 771.205057] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 771.205399] env[66641]: DEBUG nova.virt.hardware [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 771.216514] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfiguring VM instance instance-0000001c to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 771.220027] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dd2371eb-79af-4fa2-a7fb-3dd1c7d74c9a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.257833] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 771.257833] env[66641]: value = "task-5145954" [ 771.257833] env[66641]: _type = "Task" [ 771.257833] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.272400] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145954, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.314649] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145953, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141043} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.315047] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 771.315337] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 771.315583] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.328994] env[66641]: DEBUG nova.compute.manager [req-044d34f4-5f85-46cf-be16-0fbf5480615d req-22e9f22d-98c6-42b8-a4a5-d7600b5a61f6 service nova] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Received event network-vif-deleted-20765973-11e5-4b88-aa88-4c5760fa5d37 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 771.353795] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Releasing lock "refresh_cache-dad5ed23-71a1-4b55-856f-2484f8e62708" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 771.360598] env[66641]: DEBUG nova.compute.manager [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Received event network-vif-plugged-cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 771.361791] env[66641]: DEBUG oslo_concurrency.lockutils [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.361791] env[66641]: DEBUG oslo_concurrency.lockutils [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.362072] env[66641]: DEBUG oslo_concurrency.lockutils [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.362207] env[66641]: DEBUG nova.compute.manager [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] No waiting events found dispatching network-vif-plugged-cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 771.362412] env[66641]: WARNING nova.compute.manager [req-aded198a-d819-48f8-8f52-3bef37d031e1 req-e82e82fb-b155-4efa-9579-2f6330f1735d service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Received unexpected event network-vif-plugged-cdd4e9c6-89a9-4a63-8850-26be458c02fe for instance with vm_state building and task_state spawning. [ 771.464579] env[66641]: DEBUG nova.scheduler.client.report [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 771.502321] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.502541] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 771.502830] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 771.654384] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "0f6055b9-f5b4-48ba-9589-0af212808be7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.654578] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.654759] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 771.655023] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.655345] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.659440] env[66641]: INFO nova.compute.manager [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Terminating instance [ 771.662542] env[66641]: INFO nova.compute.manager [-] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Took 1.71 seconds to deallocate network for instance. [ 771.685384] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145949, 'name': CloneVM_Task, 'duration_secs': 1.691069} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.686049] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Created linked-clone VM from snapshot [ 771.687588] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6353f820-064f-4257-87af-f1fa454f05db {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.703615] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Uploading image 2b6cac88-32b1-4d94-9778-33f4fbfa2eab {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 771.743646] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 771.743646] env[66641]: value = "vm-1000674" [ 771.743646] env[66641]: _type = "VirtualMachine" [ 771.743646] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 771.744850] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-d6fd0fb0-2064-4f4e-b308-a3c69f4578ba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.757014] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lease: (returnval){ [ 771.757014] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529ebf3c-c67a-431b-2214-9ab96cbd33c9" [ 771.757014] env[66641]: _type = "HttpNfcLease" [ 771.757014] env[66641]: } obtained for exporting VM: (result){ [ 771.757014] env[66641]: value = "vm-1000674" [ 771.757014] env[66641]: _type = "VirtualMachine" [ 771.757014] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 771.758141] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the lease: (returnval){ [ 771.758141] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529ebf3c-c67a-431b-2214-9ab96cbd33c9" [ 771.758141] env[66641]: _type = "HttpNfcLease" [ 771.758141] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 771.772394] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145954, 'name': ReconfigVM_Task, 'duration_secs': 0.359111} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.774292] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfigured VM instance instance-0000001c to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 771.774816] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 771.774816] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529ebf3c-c67a-431b-2214-9ab96cbd33c9" [ 771.774816] env[66641]: _type = "HttpNfcLease" [ 771.774816] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 771.775705] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcaf3411-9349-47a2-a652-d297c0aa039f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.802750] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfiguring VM instance instance-0000001c to attach disk [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 771.803295] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-735b098d-e217-4c05-bbbc-f5e82575ea91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.827397] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 771.827397] env[66641]: value = "task-5145956" [ 771.827397] env[66641]: _type = "Task" [ 771.827397] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.839901] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145956, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.973959] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.038s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 771.976853] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.752s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 771.978742] env[66641]: INFO nova.compute.claims [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 772.007050] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.008058] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.014830] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 772.016767] env[66641]: INFO nova.scheduler.client.report [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Deleted allocations for instance d03f39a4-532b-439a-9055-19fc1e769fff [ 772.163623] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.164033] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.173021] env[66641]: DEBUG nova.compute.manager [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 772.173568] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 772.175167] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-756a3fc2-5ce8-4259-a818-74c831eb009f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.181397] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.187303] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.187303] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0fc0fbe0-30b0-40d2-aa6e-2031f06de4c2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.196760] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 772.196760] env[66641]: value = "task-5145957" [ 772.196760] env[66641]: _type = "Task" [ 772.196760] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.217457] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.274034] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 772.274034] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529ebf3c-c67a-431b-2214-9ab96cbd33c9" [ 772.274034] env[66641]: _type = "HttpNfcLease" [ 772.274034] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 772.274034] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 772.274034] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 772.288363] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 772.288363] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529ebf3c-c67a-431b-2214-9ab96cbd33c9" [ 772.288363] env[66641]: _type = "HttpNfcLease" [ 772.288363] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 772.290206] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f761ff58-9c40-4a7c-a70f-f3fb4faa6ac3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.304742] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 772.305115] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 772.397524] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145956, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.413732] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 772.414497] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 772.414926] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 772.415351] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 772.416038] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 772.416038] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 772.416227] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.416411] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 772.416735] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 772.417022] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 772.417237] env[66641]: DEBUG nova.virt.hardware [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 772.418361] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ce58234-6812-4f6c-96db-9feff550a489 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.430824] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd6cfe91-5614-4f9c-9805-fec0804533fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.454045] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 772.461130] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 772.465360] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 772.466415] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e5feb25-7751-4770-a0ef-baf57b1ed5ad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.481822] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-231f761e-1390-4bd3-bba1-025471084050 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.486497] env[66641]: DEBUG nova.network.neutron [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 772.502457] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 772.502457] env[66641]: value = "task-5145958" [ 772.502457] env[66641]: _type = "Task" [ 772.502457] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.520157] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145958, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.532408] env[66641]: DEBUG oslo_concurrency.lockutils [None req-73982633-86f0-4caa-bc56-542dbe215792 tempest-ServerShowV254Test-713230948 tempest-ServerShowV254Test-713230948-project-member] Lock "d03f39a4-532b-439a-9055-19fc1e769fff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.283s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 772.710674] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145957, 'name': PowerOffVM_Task, 'duration_secs': 0.292671} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.710674] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 772.710674] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 772.711524] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e7e6e99-457e-448c-8876-2d85f12473e2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.800683] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 772.801257] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 772.801395] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleting the datastore file [datastore2] 0f6055b9-f5b4-48ba-9589-0af212808be7 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 772.801847] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a055d337-86ab-4a09-a23d-01498d88063e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.812023] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for the task: (returnval){ [ 772.812023] env[66641]: value = "task-5145960" [ 772.812023] env[66641]: _type = "Task" [ 772.812023] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.839833] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.894445] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145956, 'name': ReconfigVM_Task, 'duration_secs': 0.692414} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.895048] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Reconfigured VM instance instance-0000001c to attach disk [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f/047a5c42-3930-4e6a-b3a5-5dbf55d44a4f.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 772.895384] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 50 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 772.906527] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 772.912839] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e23d6c16-0895-4ef8-9993-b60e5d9615a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.915033] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "40d7571f-09e2-463a-a449-36c621045819" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 772.915187] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 772.927142] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 772.927142] env[66641]: value = "task-5145961" [ 772.927142] env[66641]: _type = "Task" [ 772.927142] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.945731] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.996817] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 772.997441] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Instance network_info: |[{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 773.001835] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:c2:42', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdd4e9c6-89a9-4a63-8850-26be458c02fe', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 773.012049] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 773.012049] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 773.015543] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-234f4d7d-dbe0-43e1-ada0-32f93c034ca6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.051824] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 773.051824] env[66641]: value = "task-5145962" [ 773.051824] env[66641]: _type = "Task" [ 773.051824] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.052842] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145958, 'name': CreateVM_Task, 'duration_secs': 0.414008} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.053611] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.059249] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.059848] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.060179] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 773.060640] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bb6fec5c-ffbe-48de-8aa7-2237098218ba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.073393] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145962, 'name': CreateVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.080343] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 773.080343] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b6a394-f8cb-e260-2ec6-9b5d069528b4" [ 773.080343] env[66641]: _type = "Task" [ 773.080343] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.096035] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b6a394-f8cb-e260-2ec6-9b5d069528b4, 'name': SearchDatastore_Task, 'duration_secs': 0.0148} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.097383] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.098133] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.098241] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.098640] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.098944] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 773.100179] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1a1de03-ed6c-41ad-92a5-92f8094cf3a1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.113873] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 773.114136] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 773.114979] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-96ea7ff8-85fa-475b-881a-dc8a60feb658 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.122414] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 773.122414] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a0f32b-70b3-a73c-b329-7e4342514368" [ 773.122414] env[66641]: _type = "Task" [ 773.122414] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.138278] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a0f32b-70b3-a73c-b329-7e4342514368, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.324202] env[66641]: DEBUG oslo_vmware.api [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Task: {'id': task-5145960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.262141} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.328568] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.328888] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 773.329181] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 773.329409] env[66641]: INFO nova.compute.manager [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 773.329692] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 773.330308] env[66641]: DEBUG nova.compute.manager [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 773.330510] env[66641]: DEBUG nova.network.neutron [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 773.331130] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.331618] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.408316] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac35225-11da-425e-98ea-3a015b7b7716 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.436964] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 773.447600] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3be68cd-75fa-4d2a-8d3a-3eef62a020a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.461045] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145961, 'name': PowerOffVM_Task, 'duration_secs': 0.304668} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.476600] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.478082] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 67 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 773.484397] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4e282b-6a82-4a7b-8f10-7641a6bec921 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.517024] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5fed72-18cf-4c69-8a9d-6c8f9a23c898 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.521134] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4823967-7bf0-4f3f-b6d3-3c856bd2a990 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.537764] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40b5cc8e-5c02-4344-9b88-b10305cf3309 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.582101] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.582688] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.595205] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dbbf3e-cd8f-461d-8d30-98a032595006 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.609277] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69bb4c28-093d-44d7-8251-86c91b8f4656 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.614715] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145962, 'name': CreateVM_Task, 'duration_secs': 0.553954} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.614715] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 773.616009] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 773.616459] env[66641]: WARNING openstack [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 773.621845] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.622316] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 773.622726] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 773.634790] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57c8094b-7dd8-47d8-bb2b-20970b5230b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.638050] env[66641]: DEBUG nova.compute.provider_tree [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.649108] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 773.649108] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c9f873-3e5f-87d9-a2fd-eb7b844cabcc" [ 773.649108] env[66641]: _type = "Task" [ 773.649108] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.654367] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a0f32b-70b3-a73c-b329-7e4342514368, 'name': SearchDatastore_Task, 'duration_secs': 0.015199} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.660287] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b164e48-aa7e-44eb-9144-56d155c4a5cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.672160] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52c9f873-3e5f-87d9-a2fd-eb7b844cabcc, 'name': SearchDatastore_Task, 'duration_secs': 0.014941} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.673790] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 773.674065] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 773.674328] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.674667] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 773.674667] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522e4e12-d137-ebad-9d88-96d0c9a74907" [ 773.674667] env[66641]: _type = "Task" [ 773.674667] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.685207] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522e4e12-d137-ebad-9d88-96d0c9a74907, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.978824] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 773.991313] env[66641]: WARNING neutronclient.v2_0.client [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 774.036170] env[66641]: DEBUG nova.network.neutron [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Port 2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a binding to destination host cpu-1 is already ACTIVE {{(pid=66641) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 774.146356] env[66641]: DEBUG nova.scheduler.client.report [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 774.190698] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522e4e12-d137-ebad-9d88-96d0c9a74907, 'name': SearchDatastore_Task, 'duration_secs': 0.030279} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.190698] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 774.190921] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 774.191359] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 774.191709] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 774.192147] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3a24c09e-c890-4ff5-b8cb-8d48637977ce {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.195175] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0816f1f1-001b-447e-b783-14b46854525f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.205266] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 774.205266] env[66641]: value = "task-5145963" [ 774.205266] env[66641]: _type = "Task" [ 774.205266] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.210653] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 774.211021] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 774.212441] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-92153538-fea6-467a-a702-4c9b8f0ffcfb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.221119] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.227304] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 774.227304] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5257f034-c721-be73-6a60-61b9711f6773" [ 774.227304] env[66641]: _type = "Task" [ 774.227304] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.239157] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5257f034-c721-be73-6a60-61b9711f6773, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.436119] env[66641]: DEBUG nova.network.neutron [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 774.653945] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.676s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 774.656031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 13.515s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 774.720042] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145963, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.739027] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5257f034-c721-be73-6a60-61b9711f6773, 'name': SearchDatastore_Task, 'duration_secs': 0.013067} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.740174] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3109c704-7758-4b2d-b5f1-e81bb6fcbbfd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.748223] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 774.748223] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529e0d5e-d0fa-700a-2c79-6794e19895a7" [ 774.748223] env[66641]: _type = "Task" [ 774.748223] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.759699] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529e0d5e-d0fa-700a-2c79-6794e19895a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.939548] env[66641]: INFO nova.compute.manager [-] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Took 1.61 seconds to deallocate network for instance. [ 775.065733] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.065733] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.066046] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.159138] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "376064fc-d900-4c3d-9bc7-cc475fdc09ee" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.159493] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "376064fc-d900-4c3d-9bc7-cc475fdc09ee" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 775.164145] env[66641]: INFO nova.compute.claims [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 775.169804] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "376064fc-d900-4c3d-9bc7-cc475fdc09ee" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.010s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 775.170055] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 775.221211] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544238} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.221744] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.222099] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.222412] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21b76c57-ac71-44dd-a680-5745215dcdf3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.231543] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 775.231543] env[66641]: value = "task-5145964" [ 775.231543] env[66641]: _type = "Task" [ 775.231543] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.243917] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.261897] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529e0d5e-d0fa-700a-2c79-6794e19895a7, 'name': SearchDatastore_Task, 'duration_secs': 0.021428} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.262229] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 775.262667] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 775.262954] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39f46a73-d396-4a09-b369-e32616b66a02 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.273606] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 775.273606] env[66641]: value = "task-5145965" [ 775.273606] env[66641]: _type = "Task" [ 775.273606] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.284441] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.449827] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 775.677297] env[66641]: INFO nova.compute.resource_tracker [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating resource usage from migration b4026a95-a6cb-4f08-a135-728567b9bd8c [ 775.681428] env[66641]: DEBUG nova.compute.utils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 775.686540] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 775.686820] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 775.687179] env[66641]: WARNING neutronclient.v2_0.client [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.687641] env[66641]: WARNING neutronclient.v2_0.client [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 775.688439] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 775.688668] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 775.749612] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0788} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.753311] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.755481] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddc8fa9c-a213-4592-a9ca-059f854bde3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.782156] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.787364] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ce71c8b9-e5ad-4182-859a-b957824102f9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.825779] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145965, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531365} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.827975] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 775.828322] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 775.828779] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 775.828779] env[66641]: value = "task-5145966" [ 775.828779] env[66641]: _type = "Task" [ 775.828779] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.832904] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-58e7656e-4a5c-432c-b811-19b01c3e4331 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.852771] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145966, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.860085] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 775.860085] env[66641]: value = "task-5145967" [ 775.860085] env[66641]: _type = "Task" [ 775.860085] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.871415] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145967, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.970897] env[66641]: DEBUG nova.policy [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cb0dc64cd00a4522826f702d186c3ae8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7a6c353f022e4cf7a376a400fba55bee', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 776.072799] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 776.073344] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 776.112856] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2d5bc8-d11d-4e42-ab9b-6b288ba391eb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.122564] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4dde1f-3996-4e75-b077-637e41419883 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.155328] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95be1a9e-4676-42ee-918a-268d8f7c2a61 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.165039] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efd6ef0-d9a0-4a5f-9b05-25bc228d5a1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.181839] env[66641]: DEBUG nova.compute.provider_tree [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.200554] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 776.351233] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145966, 'name': ReconfigVM_Task, 'duration_secs': 0.470909} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.351808] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Reconfigured VM instance instance-00000024 to attach disk [datastore2] 61042df0-a727-4aa8-b2ea-bdc40899d0fc/61042df0-a727-4aa8-b2ea-bdc40899d0fc.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.352740] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aa495bf1-049f-4356-b5cd-d7e5cb8eebee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.367687] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 776.367687] env[66641]: value = "task-5145968" [ 776.367687] env[66641]: _type = "Task" [ 776.367687] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.380468] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145967, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108221} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.381630] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 776.383108] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871721a5-c8c0-41c3-a768-8d665b1dd6cf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.394157] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145968, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.421742] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 776.424136] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ebb9fccf-59dd-45c9-b86f-a9035e91d525 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.446604] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 776.446604] env[66641]: value = "task-5145969" [ 776.446604] env[66641]: _type = "Task" [ 776.446604] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.457546] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145969, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.685074] env[66641]: DEBUG nova.scheduler.client.report [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 776.730203] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Successfully created port: 9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 776.766279] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.766569] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 776.766909] env[66641]: DEBUG nova.network.neutron [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 776.878555] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145968, 'name': Rename_Task, 'duration_secs': 0.204245} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.878555] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.878835] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0d16f3ca-3d45-4168-8c70-23360f3cdb6c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.888844] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 776.888844] env[66641]: value = "task-5145970" [ 776.888844] env[66641]: _type = "Task" [ 776.888844] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.899505] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145970, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.958079] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145969, 'name': ReconfigVM_Task, 'duration_secs': 0.460002} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.958481] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.959237] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-889cad5a-97b5-4137-874f-cc080f1672a8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.968243] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 776.968243] env[66641]: value = "task-5145971" [ 776.968243] env[66641]: _type = "Task" [ 776.968243] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.982343] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145971, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.192194] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.536s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 777.192592] env[66641]: INFO nova.compute.manager [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Migrating [ 777.202817] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 13.931s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 777.215410] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 777.274467] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.274982] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.394576] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.395887] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.412470] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145970, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.477458] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 777.477458] env[66641]: WARNING openstack [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 777.493236] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145971, 'name': Rename_Task, 'duration_secs': 0.234165} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.495236] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 777.495236] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2fda7991-3f41-44a1-a4e3-05c73cd541f2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.504704] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 777.504704] env[66641]: value = "task-5145972" [ 777.504704] env[66641]: _type = "Task" [ 777.504704] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.511129] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145972, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.589023] env[66641]: DEBUG nova.network.neutron [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 777.723185] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.723354] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 777.723503] env[66641]: DEBUG nova.network.neutron [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 777.899798] env[66641]: DEBUG oslo_vmware.api [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145970, 'name': PowerOnVM_Task, 'duration_secs': 0.612476} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.899968] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 777.900118] env[66641]: DEBUG nova.compute.manager [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 777.900977] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2b8649-4eca-4345-b6ab-bead87b62b9b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.013068] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145972, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.092532] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 778.225820] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Applying migration context for instance 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f as it has an incoming, in-progress migration 553a2d88-0488-434e-be83-aa85f338f673. Migration status is post-migrating {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 778.226963] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Applying migration context for instance 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 as it has an incoming, in-progress migration b4026a95-a6cb-4f08-a135-728567b9bd8c. Migration status is pre-migrating {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1046}} [ 778.228660] env[66641]: INFO nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating resource usage from migration 553a2d88-0488-434e-be83-aa85f338f673 [ 778.229031] env[66641]: INFO nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating resource usage from migration b4026a95-a6cb-4f08-a135-728567b9bd8c [ 778.232131] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.232541] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.254513] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.254941] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance bc13bbe2-6e02-4d98-9e50-94f772d89ac0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.255197] env[66641]: WARNING nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 8235441c-c046-49da-a5a5-92bd85e17982 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 778.256999] env[66641]: WARNING nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 530f1e0e-5911-434e-bf20-edfd9778d7ad is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 778.256999] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance b9d032da-031e-42e0-86e2-95254c1ceac1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.256999] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance a7575399-f72d-4d2c-b57b-f2f2f591c8a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.256999] env[66641]: WARNING nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 0f6055b9-f5b4-48ba-9589-0af212808be7 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 778.257204] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance e508b1e9-cbb0-408c-98e1-d54ceaec1f1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257204] env[66641]: WARNING nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 25c7bd59-ec24-4d30-840b-3c4549dbf669 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 778.257204] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 05882781-78be-4568-95f4-2fccc4cf4dfe actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257204] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance dad5ed23-71a1-4b55-856f-2484f8e62708 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257336] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 61042df0-a727-4aa8-b2ea-bdc40899d0fc actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257336] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 081457ae-e152-410c-bca7-4d43b95eee10 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257336] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Migration 553a2d88-0488-434e-be83-aa85f338f673 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 778.257336] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257450] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 48f99287-b737-45fa-ad59-9e1425afa3d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257450] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.257450] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Migration b4026a95-a6cb-4f08-a135-728567b9bd8c is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1745}} [ 778.257450] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 778.321793] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 778.323662] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-425083c9-8c8f-4561-9bf7-fe92d61ef5ba {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.334489] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 778.334794] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 778.335021] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 778.335264] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 778.335435] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 778.335641] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 778.335876] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.336069] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 778.336268] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 778.336494] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 778.337753] env[66641]: DEBUG nova.virt.hardware [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 778.337753] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 778.337753] env[66641]: value = "task-5145973" [ 778.337753] env[66641]: _type = "Task" [ 778.337753] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.340350] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31de3a8-c7bb-46c7-b050-5cd2b90b7cfa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.357567] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 778.358590] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee4dabe0-9a4e-466b-aef5-834aba0d68a4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.365482] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f121bd44-7ba5-408e-a0b3-a10a0f53af74 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.377907] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 778.377907] env[66641]: ERROR oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk due to incomplete transfer. [ 778.378940] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f6ea8106-4dc0-40d4-b862-74ad22104e97 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.381668] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] VM already powered off {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 778.381668] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 778.381668] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.381668] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.381875] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 778.391471] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a079f2b-eb3e-486e-a84f-cc89e59aae42 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.397629] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.398152] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.414510] env[66641]: DEBUG oslo_vmware.rw_handles [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5258a900-2910-f0c7-ef96-cce504f24c16/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 778.414510] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Uploaded image a21499e1-a4d6-4fad-b53b-341f9044cc53 to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 778.418808] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 778.424691] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5cdcd415-a03a-422b-aff1-fc712dc5dc96 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.427060] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 778.427355] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 778.438652] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d1a73d56-427d-41e3-99e2-94b4186ba54f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.442152] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 778.453487] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Successfully updated port: 9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 778.459322] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 778.459322] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527dc6a6-edb3-5665-2b4c-c26427fb1ad9" [ 778.459322] env[66641]: _type = "Task" [ 778.459322] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.459920] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 778.459920] env[66641]: value = "task-5145974" [ 778.459920] env[66641]: _type = "Task" [ 778.459920] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.481067] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145974, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.481585] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527dc6a6-edb3-5665-2b4c-c26427fb1ad9, 'name': SearchDatastore_Task, 'duration_secs': 0.013601} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.482418] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd2c7dc0-cbfe-4e36-9d3e-5a4209d26b42 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.490710] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 778.490710] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fa8850-044b-0437-d31f-5f9b01b00019" [ 778.490710] env[66641]: _type = "Task" [ 778.490710] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.501961] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fa8850-044b-0437-d31f-5f9b01b00019, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.504335] env[66641]: DEBUG nova.compute.manager [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Received event network-changed-cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 778.504677] env[66641]: DEBUG nova.compute.manager [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Refreshing instance network info cache due to event network-changed-cdd4e9c6-89a9-4a63-8850-26be458c02fe. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 778.504677] env[66641]: DEBUG oslo_concurrency.lockutils [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Acquiring lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.504899] env[66641]: DEBUG oslo_concurrency.lockutils [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Acquired lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.504984] env[66641]: DEBUG nova.network.neutron [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Refreshing network info cache for port cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 778.509688] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.509688] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.522987] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.523500] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.538142] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145972, 'name': PowerOnVM_Task} progress is 90%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.633732] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef5df2bf-d015-4c91-86f4-32bc64376be7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.657918] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975641bc-1bbd-4782-b40c-fe3d7c03de81 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.669545] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 83 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 778.681759] env[66641]: DEBUG nova.network.neutron [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 778.694246] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.694324] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.761223] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 207483d3-803c-495b-9b93-6f986f3ca56e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 778.794233] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 778.794412] env[66641]: WARNING openstack [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 778.819835] env[66641]: DEBUG nova.compute.manager [req-6fa36206-caf2-4145-b7a1-a7b1914dac79 req-337d7d04-bb5d-4539-874b-bc6ef671c9f3 service nova] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Received event network-vif-deleted-298db8ec-67aa-4b6a-8981-4f1a72602ef3 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 778.895884] env[66641]: DEBUG nova.network.neutron [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updated VIF entry in instance network info cache for port cdd4e9c6-89a9-4a63-8850-26be458c02fe. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 778.896330] env[66641]: DEBUG nova.network.neutron [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 778.949166] env[66641]: INFO nova.compute.manager [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Rebuilding instance [ 778.963477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.963477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquired lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 778.963477] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 778.972151] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145974, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.013380] env[66641]: DEBUG nova.compute.manager [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 779.020963] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80b28ee8-99bc-47a6-8a37-57b654138644 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.025637] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fa8850-044b-0437-d31f-5f9b01b00019, 'name': SearchDatastore_Task, 'duration_secs': 0.012766} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.026659] env[66641]: DEBUG oslo_concurrency.lockutils [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.027061] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/cd910b37-6707-4868-b172-79fffc590a51-rescue.vmdk. {{(pid=66641) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 779.028330] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50069100-834a-4546-ac1a-418a2da5486a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.043993] env[66641]: DEBUG oslo_vmware.api [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5145972, 'name': PowerOnVM_Task, 'duration_secs': 1.086679} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.045497] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 779.046115] env[66641]: INFO nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Took 9.15 seconds to spawn the instance on the hypervisor. [ 779.046760] env[66641]: DEBUG nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 779.048578] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b4fd641-6fcc-4e1f-a874-4655400915d3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.057046] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 779.057046] env[66641]: value = "task-5145975" [ 779.057046] env[66641]: _type = "Task" [ 779.057046] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.077392] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145975, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.176900] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 779.177394] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56f87c4f-fe04-4d70-a12b-bf31b88cf941 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.184468] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.189777] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 779.189777] env[66641]: value = "task-5145976" [ 779.189777] env[66641]: _type = "Task" [ 779.189777] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.203922] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145976, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.264317] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 40d7571f-09e2-463a-a449-36c621045819 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1797}} [ 779.264639] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 15 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 779.266365] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3584MB phys_disk=100GB used_disk=15GB total_vcpus=48 used_vcpus=15 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '13', 'num_vm_active': '11', 'num_task_None': '5', 'num_os_type_None': '13', 'num_proj_6624f269bb8840e98c10259f006ce962': '2', 'io_workload': '4', 'num_proj_2ab82c335b15478788a818319bf11c03': '1', 'num_proj_04f038652b90481f9629847d20f93f84': '1', 'num_task_resize_migrated': '1', 'num_proj_8449c95dd0f74492929e4d1d492ac9d2': '1', 'num_task_image_uploading': '2', 'num_proj_7ecf362f0f454995adbf560fa17e1dab': '2', 'num_task_resize_prep': '1', 'num_proj_17792883c6504d0f9f91fb6beb25b087': '1', 'num_task_rescuing': '1', 'num_proj_b740816edb13463cb24e1caac7dbd569': '1', 'num_task_rebuild_spawning': '1', 'num_proj_9c7f8a6e5b2c4c9593f4d2eaf86eb5ac': '1', 'num_proj_076797f59df643b4b30fc26b60e2189b': '1', 'num_vm_building': '2', 'num_task_spawning': '2', 'num_proj_b2c8481015524aee95a933f61082faec': '1', 'num_proj_7a6c353f022e4cf7a376a400fba55bee': '1'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 779.400497] env[66641]: DEBUG oslo_concurrency.lockutils [req-96499a89-6db0-4cdd-8155-ad364354ef70 req-7d0c2ffe-6002-42ac-abc3-9c854b0bb7db service nova] Releasing lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 779.468244] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.468244] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.475929] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 779.491818] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145974, 'name': Destroy_Task, 'duration_secs': 0.644594} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.492216] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Destroyed the VM [ 779.492474] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 779.492806] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-c38d1d90-528d-4501-ac59-dd39b02381ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.509715] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 779.509715] env[66641]: value = "task-5145977" [ 779.509715] env[66641]: _type = "Task" [ 779.509715] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.519843] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145977, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.583239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "b9d032da-031e-42e0-86e2-95254c1ceac1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.583239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.583239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.583239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.583808] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.590763] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145975, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.591402] env[66641]: INFO nova.compute.manager [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Took 23.35 seconds to build instance. [ 779.592753] env[66641]: INFO nova.compute.manager [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Terminating instance [ 779.698070] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 779.698625] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 779.718946] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145976, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.780155] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6576e92-389a-4eb7-a79b-83f61bda300c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.788181] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77561347-487b-4d2b-957c-6185c61e36f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.821554] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62450018-ceb3-4345-948c-439544ccb1b4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.831211] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab7fb26-6056-45c1-9fc0-103686927dbb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.846976] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.847621] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.847621] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 779.847621] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 779.847792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 779.849537] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 779.851124] env[66641]: INFO nova.compute.manager [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Terminating instance [ 780.019813] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145977, 'name': RemoveSnapshot_Task} progress is 50%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.052907] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.053480] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-49c8e8c4-dc34-4537-b97f-bf5786a1a6e2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.062049] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 780.062049] env[66641]: value = "task-5145978" [ 780.062049] env[66641]: _type = "Task" [ 780.062049] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.072320] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145975, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.578132} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.073249] env[66641]: INFO nova.virt.vmwareapi.ds_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/cd910b37-6707-4868-b172-79fffc590a51-rescue.vmdk. [ 780.074437] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1eaa70b-6e5e-4af6-88c1-ab861433f1e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.082658] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145978, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.108158] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/cd910b37-6707-4868-b172-79fffc590a51-rescue.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 780.108506] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1b32a234-b7b8-4683-bf40-28deaa9970cc tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.893s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.109186] env[66641]: DEBUG nova.compute.manager [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 780.109412] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.109917] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b95e4d89-a9e8-464a-965c-874430a51df9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.124846] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b2d571-6288-4760-ad49-583f705190e8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.135115] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 780.136692] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f696d55d-0723-4d1b-a05e-457cc82a2244 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.139181] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 780.139181] env[66641]: value = "task-5145979" [ 780.139181] env[66641]: _type = "Task" [ 780.139181] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.146025] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 780.146025] env[66641]: value = "task-5145980" [ 780.146025] env[66641]: _type = "Task" [ 780.146025] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.150349] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145979, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.161943] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145980, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.207685] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.208117] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.228581] env[66641]: DEBUG oslo_vmware.api [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5145976, 'name': PowerOnVM_Task, 'duration_secs': 0.554806} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.229027] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 780.229283] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c943d1b1-c05c-46e3-9344-0ef86fffdbd4 tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance '047a5c42-3930-4e6a-b3a5-5dbf55d44a4f' progress to 100 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 780.334052] env[66641]: DEBUG nova.network.neutron [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Updating instance_info_cache with network_info: [{"id": "9738572a-f121-4d8c-be4f-771175b83bed", "address": "fa:16:3e:52:f5:5b", "network": {"id": "d85cd610-7e1a-401c-af14-1b382b4b12c8", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1986360916-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a6c353f022e4cf7a376a400fba55bee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9738572a-f1", "ovs_interfaceid": "9738572a-f121-4d8c-be4f-771175b83bed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 780.358449] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "refresh_cache-61042df0-a727-4aa8-b2ea-bdc40899d0fc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.358705] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquired lock "refresh_cache-61042df0-a727-4aa8-b2ea-bdc40899d0fc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.358885] env[66641]: DEBUG nova.network.neutron [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 780.379298] env[66641]: ERROR nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [req-be150120-f93d-4ae7-bfe1-7f4d682fb26a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 94, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 750ffd2d-5e46-4240-a614-995f2be7c9cb. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-be150120-f93d-4ae7-bfe1-7f4d682fb26a"}]} [ 780.407716] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Refreshing inventories for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:822}} [ 780.423337] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Updating ProviderTree inventory for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:786}} [ 780.423569] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Updating inventory in ProviderTree for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.436165] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Refreshing aggregate associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, aggregates: None {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:831}} [ 780.457683] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Refreshing trait associations for resource provider 750ffd2d-5e46-4240-a614-995f2be7c9cb, traits: HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=66641) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:843}} [ 780.523352] env[66641]: DEBUG oslo_vmware.api [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5145977, 'name': RemoveSnapshot_Task, 'duration_secs': 0.907795} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.523352] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 780.523352] env[66641]: INFO nova.compute.manager [None req-a004132a-e60b-4dc6-a0d7-7958700e7c1f tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 15.70 seconds to snapshot the instance on the hypervisor. [ 780.576156] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145978, 'name': PowerOffVM_Task, 'duration_secs': 0.2372} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.576270] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.576942] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 780.580842] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b515da0f-cd27-4e01-b545-ba40b9681079 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.589477] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.589762] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-388a1e7b-0791-4990-92cf-f27e4fcb5984 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.621118] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.621453] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.621689] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Deleting the datastore file [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.622055] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-793cd952-b6e2-47ea-8674-e06ed85345d7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.635738] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 780.635738] env[66641]: value = "task-5145982" [ 780.635738] env[66641]: _type = "Task" [ 780.635738] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.657980] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145982, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.658295] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145979, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.672949] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145980, 'name': PowerOffVM_Task, 'duration_secs': 0.261537} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.673369] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 780.673545] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 780.673830] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59e263de-70a3-41fe-8f37-09c981edfd2d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.726996] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3611e1c-5e42-40c4-93e3-e05b92b49f58 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.753774] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 0 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 780.763737] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 780.764109] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 780.764494] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Deleting the datastore file [datastore1] b9d032da-031e-42e0-86e2-95254c1ceac1 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 780.765984] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8d3dbace-15dc-4aaa-92f7-4d0d20ee13e6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.775597] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for the task: (returnval){ [ 780.775597] env[66641]: value = "task-5145984" [ 780.775597] env[66641]: _type = "Task" [ 780.775597] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.789031] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.831627] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2635e23-b0bb-4600-9e21-8b7686ac44df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.836992] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Releasing lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 780.837801] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Instance network_info: |[{"id": "9738572a-f121-4d8c-be4f-771175b83bed", "address": "fa:16:3e:52:f5:5b", "network": {"id": "d85cd610-7e1a-401c-af14-1b382b4b12c8", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1986360916-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a6c353f022e4cf7a376a400fba55bee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9738572a-f1", "ovs_interfaceid": "9738572a-f121-4d8c-be4f-771175b83bed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 780.838077] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:f5:5b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69076131-87ac-46dd-9d5d-8d1b4ea7dec6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9738572a-f121-4d8c-be4f-771175b83bed', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 780.848031] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Creating folder: Project (7a6c353f022e4cf7a376a400fba55bee). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 780.849231] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-67addd01-149c-407a-beec-f306104fd0c1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.852064] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9501f592-5550-4659-84d7-1935dbd85794 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.888945] env[66641]: WARNING openstack [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 780.889362] env[66641]: WARNING openstack [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 780.894612] env[66641]: DEBUG nova.network.neutron [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 780.899012] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f514fa-b05c-4ca6-ad2a-ef446557f49c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.901809] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Created folder: Project (7a6c353f022e4cf7a376a400fba55bee) in parent group-v1000566. [ 780.902360] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Creating folder: Instances. Parent ref: group-v1000677. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 780.902623] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35f28797-dafb-42bc-9b88-856aad58bd1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.911938] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22592b3c-77fe-4405-b439-64c1dd9ab5b3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.917627] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Created folder: Instances in parent group-v1000677. [ 780.917895] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 780.918551] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 780.918770] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68bb937b-da9c-4e2a-9ee4-6c516c236479 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.945361] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.950292] env[66641]: DEBUG nova.compute.manager [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Received event network-vif-plugged-9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 780.950448] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Acquiring lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 780.950670] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 780.950846] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 780.951039] env[66641]: DEBUG nova.compute.manager [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] No waiting events found dispatching network-vif-plugged-9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 780.951206] env[66641]: WARNING nova.compute.manager [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Received unexpected event network-vif-plugged-9738572a-f121-4d8c-be4f-771175b83bed for instance with vm_state building and task_state spawning. [ 780.951360] env[66641]: DEBUG nova.compute.manager [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Received event network-changed-9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 780.951506] env[66641]: DEBUG nova.compute.manager [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Refreshing instance network info cache due to event network-changed-9738572a-f121-4d8c-be4f-771175b83bed. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 780.951680] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Acquiring lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.951807] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Acquired lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 780.951957] env[66641]: DEBUG nova.network.neutron [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Refreshing network info cache for port 9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 780.955106] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 780.955106] env[66641]: value = "task-5145987" [ 780.955106] env[66641]: _type = "Task" [ 780.955106] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.968075] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145987, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.998271] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 780.999300] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b241493-18a3-4ae1-9611-b7300959a81b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.007029] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 781.007207] env[66641]: ERROR oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk due to incomplete transfer. [ 781.007453] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-49c8fdac-5b10-4018-b6bb-e53559336033 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.015712] env[66641]: DEBUG oslo_vmware.rw_handles [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5210b2f1-e069-d5c2-a2fa-1d61243bce13/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 781.015915] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Uploaded image 2b6cac88-32b1-4d94-9778-33f4fbfa2eab to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 781.017374] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 781.020594] env[66641]: DEBUG nova.network.neutron [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 781.021874] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-dd8b07af-f19f-40b1-811f-36dd79c03682 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.031511] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 781.031511] env[66641]: value = "task-5145988" [ 781.031511] env[66641]: _type = "Task" [ 781.031511] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.045863] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145988, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.150626] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5145982, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132925} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.154118] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.154367] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.154625] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.157485] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145979, 'name': ReconfigVM_Task, 'duration_secs': 0.668245} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.157774] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Reconfigured VM instance instance-00000023 to attach disk [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708/cd910b37-6707-4868-b172-79fffc590a51-rescue.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 781.158736] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc38fa3-af7a-44d3-961a-25a0e8b8a37c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.192330] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87c22c52-ccc1-4ed8-a67c-33bdd118fb68 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.211582] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 781.211582] env[66641]: value = "task-5145989" [ 781.211582] env[66641]: _type = "Task" [ 781.211582] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.223057] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145989, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.269212] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.270107] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-25d86c76-d047-471d-baa9-2faf32eef886 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.281742] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 781.281742] env[66641]: value = "task-5145990" [ 781.281742] env[66641]: _type = "Task" [ 781.281742] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.299640] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145990, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.307023] env[66641]: DEBUG oslo_vmware.api [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Task: {'id': task-5145984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158657} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.307023] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 781.307023] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 781.307023] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 781.307023] env[66641]: INFO nova.compute.manager [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Took 1.20 seconds to destroy the instance on the hypervisor. [ 781.307351] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 781.307351] env[66641]: DEBUG nova.compute.manager [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 781.307351] env[66641]: DEBUG nova.network.neutron [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 781.307351] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.307351] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.355943] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.357401] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.450986] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 781.458676] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.458676] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.486861] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145987, 'name': CreateVM_Task, 'duration_secs': 0.42513} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.486861] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 781.486861] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.486861] env[66641]: WARNING openstack [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.492911] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.493408] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.493974] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 781.494922] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab79c199-3d5e-47f7-8e8d-d6de7353d792 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.502898] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 781.502898] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bc8f98-dce4-743c-4293-7f97afda3be8" [ 781.502898] env[66641]: _type = "Task" [ 781.502898] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.519035] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bc8f98-dce4-743c-4293-7f97afda3be8, 'name': SearchDatastore_Task, 'duration_secs': 0.012359} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.519035] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.519800] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 781.519800] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.519800] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 781.520022] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 781.520361] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b967f903-6ccb-4505-a2fc-513e5f5b2e9c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.527707] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Releasing lock "refresh_cache-61042df0-a727-4aa8-b2ea-bdc40899d0fc" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 781.528199] env[66641]: DEBUG nova.compute.manager [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 781.528399] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.529674] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b20536-d220-4877-9cef-8f7db504d325 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.536157] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 781.536462] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 781.541489] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a7a128e-25ba-47c6-815c-ecfece29046d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.548482] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.549708] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1947900-d79b-4a50-a612-b79a96f72238 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.553316] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 781.553316] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528cc56c-1208-6133-688d-d493ad10a781" [ 781.553316] env[66641]: _type = "Task" [ 781.553316] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.559103] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145988, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.569027] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 781.569027] env[66641]: value = "task-5145991" [ 781.569027] env[66641]: _type = "Task" [ 781.569027] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.575526] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528cc56c-1208-6133-688d-d493ad10a781, 'name': SearchDatastore_Task, 'duration_secs': 0.013442} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.576887] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9a4d662d-54ec-4444-b570-93b380e165ad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.582788] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145991, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.587840] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 781.587840] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5267c49e-b5a3-18eb-ca6f-dc784439af5c" [ 781.587840] env[66641]: _type = "Task" [ 781.587840] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.597474] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5267c49e-b5a3-18eb-ca6f-dc784439af5c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.727514] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145989, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.746325] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.746762] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.800600] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145990, 'name': PowerOffVM_Task, 'duration_secs': 0.289181} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.800907] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.801143] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 17 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 781.826898] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 781.827299] env[66641]: WARNING openstack [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 781.938185] env[66641]: DEBUG nova.network.neutron [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Updated VIF entry in instance network info cache for port 9738572a-f121-4d8c-be4f-771175b83bed. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 781.938562] env[66641]: DEBUG nova.network.neutron [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Updating instance_info_cache with network_info: [{"id": "9738572a-f121-4d8c-be4f-771175b83bed", "address": "fa:16:3e:52:f5:5b", "network": {"id": "d85cd610-7e1a-401c-af14-1b382b4b12c8", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1986360916-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7a6c353f022e4cf7a376a400fba55bee", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69076131-87ac-46dd-9d5d-8d1b4ea7dec6", "external-id": "nsx-vlan-transportzone-327", "segmentation_id": 327, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9738572a-f1", "ovs_interfaceid": "9738572a-f121-4d8c-be4f-771175b83bed", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 781.955255] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 781.955642] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.753s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.956261] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.664s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.956443] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.958850] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.980s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.959061] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 781.963906] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.461s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 781.965245] env[66641]: INFO nova.compute.claims [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.005689] env[66641]: INFO nova.scheduler.client.report [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleted allocations for instance 530f1e0e-5911-434e-bf20-edfd9778d7ad [ 782.011435] env[66641]: INFO nova.scheduler.client.report [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Deleted allocations for instance 8235441c-c046-49da-a5a5-92bd85e17982 [ 782.052029] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145988, 'name': Destroy_Task, 'duration_secs': 0.642} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.052029] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Destroyed the VM [ 782.052029] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 782.052029] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cbb2e275-ce7c-4d6c-8adc-e2619e1f8dd0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.061479] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 782.061479] env[66641]: value = "task-5145992" [ 782.061479] env[66641]: _type = "Task" [ 782.061479] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.079055] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145992, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.083633] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145991, 'name': PowerOffVM_Task, 'duration_secs': 0.476057} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.083840] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 782.084012] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 782.084572] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-77cfc9ac-d703-4376-8451-29cd7bd3a5aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.100333] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5267c49e-b5a3-18eb-ca6f-dc784439af5c, 'name': SearchDatastore_Task, 'duration_secs': 0.011772} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.100333] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.102268] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db/34a98372-2ab7-4b21-8a0e-2fc3b91ef4db.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 782.102268] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d552e304-eac6-412e-88e7-3fc836dab593 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.110740] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 782.110740] env[66641]: value = "task-5145994" [ 782.110740] env[66641]: _type = "Task" [ 782.110740] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.115843] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 782.116075] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 782.116255] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Deleting the datastore file [datastore2] 61042df0-a727-4aa8-b2ea-bdc40899d0fc {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 782.116974] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-67c23ea3-3b73-47ec-b8f8-a3ee17e42a33 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.125953] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145994, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.128248] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for the task: (returnval){ [ 782.128248] env[66641]: value = "task-5145995" [ 782.128248] env[66641]: _type = "Task" [ 782.128248] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.137330] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 782.137567] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 782.142367] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145995, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.199438] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 782.199695] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 782.199852] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 782.200044] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 782.200226] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 782.200399] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 782.200618] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.200791] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 782.200959] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 782.201141] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 782.201313] env[66641]: DEBUG nova.virt.hardware [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 782.202327] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3422c9a-4df0-4e41-a1f7-28e0e355c33d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.212457] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f113b26-0125-40af-9807-13b624845d9c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.229241] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145989, 'name': ReconfigVM_Task, 'duration_secs': 0.8334} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.237539] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 782.238074] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.244185] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.244568] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78142907-36bd-4b69-b49e-6a3054183e1a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.247123] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 782.247123] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7dddfd00-2fb2-4ce5-b1cc-961955b137ab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.265771] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 782.265771] env[66641]: value = "task-5145996" [ 782.265771] env[66641]: _type = "Task" [ 782.265771] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.267037] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.267037] env[66641]: value = "task-5145997" [ 782.267037] env[66641]: _type = "Task" [ 782.267037] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.287258] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145997, 'name': CreateVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.289882] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145996, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.310611] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 782.310964] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 782.311082] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 782.311272] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 782.311430] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 782.311677] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 782.311966] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 782.312162] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 782.312415] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 782.312660] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 782.312904] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 782.318488] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89a437d3-9cd6-485c-8f3d-716333cbe280 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.340692] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 782.340692] env[66641]: value = "task-5145998" [ 782.340692] env[66641]: _type = "Task" [ 782.340692] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.355671] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145998, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.442100] env[66641]: DEBUG oslo_concurrency.lockutils [req-00ce6893-665a-477b-8c9c-9edd871f49b8 req-fbc5d9ff-efc5-4798-8f98-9f0b3a9a879a service nova] Releasing lock "refresh_cache-34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 782.442933] env[66641]: DEBUG nova.compute.manager [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Stashing vm_state: active {{(pid=66641) _prep_resize /opt/stack/nova/nova/compute/manager.py:6193}} [ 782.517406] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b01648c1-8e9c-4911-bcfb-f76a254ebfce tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "530f1e0e-5911-434e-bf20-edfd9778d7ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.554s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.528884] env[66641]: DEBUG oslo_concurrency.lockutils [None req-23bfb58e-7777-48f0-b9ed-af7406dcc549 tempest-ServerMetadataNegativeTestJSON-207852544 tempest-ServerMetadataNegativeTestJSON-207852544-project-member] Lock "8235441c-c046-49da-a5a5-92bd85e17982" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.961s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 782.581971] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145992, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.626236] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145994, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.646474] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 782.649465] env[66641]: DEBUG oslo_vmware.api [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Task: {'id': task-5145995, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.125982} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.649742] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.649930] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.650138] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.650479] env[66641]: INFO nova.compute.manager [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Took 1.12 seconds to destroy the instance on the hypervisor. [ 782.650721] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 782.651490] env[66641]: DEBUG nova.compute.manager [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 782.651638] env[66641]: DEBUG nova.network.neutron [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 782.652306] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.652844] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.701059] env[66641]: DEBUG nova.network.neutron [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 782.703078] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 782.703777] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 782.766940] env[66641]: DEBUG nova.compute.manager [req-8ef09587-9c06-4491-8ff9-c735c9769368 req-55faa9e7-ae6f-45da-857a-0f96507a4109 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Received event network-vif-deleted-0b91b13d-bfce-46a3-acc4-217925393f9a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 782.767287] env[66641]: INFO nova.compute.manager [req-8ef09587-9c06-4491-8ff9-c735c9769368 req-55faa9e7-ae6f-45da-857a-0f96507a4109 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Neutron deleted interface 0b91b13d-bfce-46a3-acc4-217925393f9a; detaching it from the instance and deleting it from the info cache [ 782.767639] env[66641]: DEBUG nova.network.neutron [req-8ef09587-9c06-4491-8ff9-c735c9769368 req-55faa9e7-ae6f-45da-857a-0f96507a4109 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.791495] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145996, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.798689] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145997, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.851195] env[66641]: WARNING oslo_messaging._drivers.amqpdriver [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 782.855658] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5145998, 'name': ReconfigVM_Task, 'duration_secs': 0.314579} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.856591] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 33 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 782.923484] env[66641]: DEBUG nova.compute.manager [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 782.925528] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cebfc86f-20bf-4714-9d07-e480e4587fb1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.930425] env[66641]: DEBUG nova.network.neutron [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 782.973909] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.073534] env[66641]: DEBUG oslo_vmware.api [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5145992, 'name': RemoveSnapshot_Task, 'duration_secs': 0.95574} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.076396] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 783.076590] env[66641]: INFO nova.compute.manager [None req-85446847-17d9-4451-9cec-55bc1af245b5 tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Took 15.58 seconds to snapshot the instance on the hypervisor. [ 783.084875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.084875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.084875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.084875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.085160] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 783.086943] env[66641]: INFO nova.compute.manager [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Terminating instance [ 783.123152] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145994, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.619889} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.127040] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db/34a98372-2ab7-4b21-8a0e-2fc3b91ef4db.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 783.127177] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 783.128273] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9556ca75-2260-4e90-9b64-60e339b3cb52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.136374] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 783.136374] env[66641]: value = "task-5145999" [ 783.136374] env[66641]: _type = "Task" [ 783.136374] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.148993] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.178040] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.214717] env[66641]: DEBUG nova.network.neutron [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 783.285127] env[66641]: DEBUG oslo_vmware.api [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5145996, 'name': PowerOnVM_Task, 'duration_secs': 0.585573} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.285392] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fa871f68-7abb-4029-a2cf-db2ca506a18f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.291286] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 783.294757] env[66641]: DEBUG nova.compute.manager [None req-dd8a2866-38dd-4347-ad48-644c03c045a4 tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 783.295613] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-515e96c7-068b-4a1f-8ecc-275c23628790 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.305890] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5145997, 'name': CreateVM_Task, 'duration_secs': 0.661225} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.310658] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d0e005-fee4-4e0f-a4ea-e6ef5d3c3b5e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.324071] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 783.328817] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.328992] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.329334] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 783.332606] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39847389-5f76-4a29-8dac-484828a5a73c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.340107] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 783.340107] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523042b7-0c0d-3539-2c00-0d38d7e603a0" [ 783.340107] env[66641]: _type = "Task" [ 783.340107] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.359957] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 783.360235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 783.360424] env[66641]: DEBUG nova.compute.manager [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Going to confirm migration 2 {{(pid=66641) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 783.361918] env[66641]: DEBUG nova.compute.manager [req-8ef09587-9c06-4491-8ff9-c735c9769368 req-55faa9e7-ae6f-45da-857a-0f96507a4109 service nova] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Detach interface failed, port_id=0b91b13d-bfce-46a3-acc4-217925393f9a, reason: Instance b9d032da-031e-42e0-86e2-95254c1ceac1 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 783.367734] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:51:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ed028205-0b7e-4e53-9191-4b8a2662b934',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-192737414',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 783.368055] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 783.368289] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 783.368454] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 783.368597] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 783.368742] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 783.368942] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 783.369107] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 783.369299] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 783.369467] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 783.369634] env[66641]: DEBUG nova.virt.hardware [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 783.375398] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfiguring VM instance instance-0000001e to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 783.380740] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b342385-7c3a-4ecd-93e2-80adcc69a446 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.395636] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]523042b7-0c0d-3539-2c00-0d38d7e603a0, 'name': SearchDatastore_Task, 'duration_secs': 0.012067} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.402834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.403132] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.403373] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.403529] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.403705] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.406034] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-afdf5797-70e4-4527-ab27-409840d087d7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.411105] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 783.411105] env[66641]: value = "task-5146000" [ 783.411105] env[66641]: _type = "Task" [ 783.411105] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.417855] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.418063] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 783.425033] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1f0faae8-0e04-4f0b-aeb7-3497efd7011a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.427431] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146000, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.432042] env[66641]: INFO nova.compute.manager [-] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Took 2.13 seconds to deallocate network for instance. [ 783.432472] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 783.432472] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528b61c5-cfa9-5091-8a58-d90ab63923fa" [ 783.432472] env[66641]: _type = "Task" [ 783.432472] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.443788] env[66641]: INFO nova.compute.manager [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] instance snapshotting [ 783.449778] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ac52f1-f14e-4802-8f1f-0e2e19ea07b2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.457608] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]528b61c5-cfa9-5091-8a58-d90ab63923fa, 'name': SearchDatastore_Task, 'duration_secs': 0.012174} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.460149] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74ca396b-5b9b-4db5-8079-b8dab4981de7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.477696] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9443ca29-922b-41ac-86f5-af55b710a55a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.482864] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 783.482864] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5282aba5-e43f-e17f-7e03-c130cac13b5f" [ 783.482864] env[66641]: _type = "Task" [ 783.482864] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.500156] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5282aba5-e43f-e17f-7e03-c130cac13b5f, 'name': SearchDatastore_Task, 'duration_secs': 0.010614} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.500156] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 783.500156] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 783.500156] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98c2c3ba-e955-4461-ad98-3a046fa6100a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.503874] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3b1d143-d6e4-439f-b112-6d189a4cb787 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.514562] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c85070-74a5-4315-adf0-0cf0c3af5634 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.519655] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 783.519655] env[66641]: value = "task-5146001" [ 783.519655] env[66641]: _type = "Task" [ 783.519655] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.554760] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0089e29d-dc7b-4400-838d-13aa01a63261 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.567797] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.576320] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e49ccb-d01d-46ec-bb8f-78484e7c327d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.592135] env[66641]: DEBUG nova.compute.manager [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 783.592394] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.592914] env[66641]: DEBUG nova.compute.provider_tree [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.595671] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3177a0e4-8c39-4086-9c54-c325dfdf388a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.605457] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 783.605814] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58f1fbae-427f-4bfc-8d78-2e802767d129 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.616484] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 783.616484] env[66641]: value = "task-5146002" [ 783.616484] env[66641]: _type = "Task" [ 783.616484] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.629797] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5146002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.649911] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145999, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.720297] env[66641]: INFO nova.compute.manager [-] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Took 1.07 seconds to deallocate network for instance. [ 783.872804] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.872804] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.916232] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 783.916746] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 783.923083] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.923227] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquired lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 783.923660] env[66641]: DEBUG nova.network.neutron [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 783.923660] env[66641]: DEBUG nova.objects.instance [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lazy-loading 'info_cache' on Instance uuid 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 784.754633] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.755726] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 784.756106] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "05882781-78be-4568-95f4-2fccc4cf4dfe" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.756323] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.756509] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.756676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.756881] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 784.759246] env[66641]: DEBUG nova.scheduler.client.report [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 784.763033] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.763986] env[66641]: INFO nova.compute.manager [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Terminating instance [ 784.769291] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d47e09ed-d258-4b32-8887-64577e049b40 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.798258] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146001, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.97668} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.806863] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 784.807254] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 784.807738] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 784.807738] env[66641]: value = "task-5146003" [ 784.807738] env[66641]: _type = "Task" [ 784.807738] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.808093] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5146002, 'name': PowerOffVM_Task, 'duration_secs': 0.510275} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.809407] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146000, 'name': ReconfigVM_Task, 'duration_secs': 0.346657} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.810381] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a347cadd-34c7-449b-9fd5-e35c4da7565d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.814336] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 784.814336] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 784.815035] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfigured VM instance instance-0000001e to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 784.822794] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f488eb26-bcf8-42bc-812d-3c4d0f00989e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.824880] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3adb6a1-13be-4fe1-8a7e-cc13a0e8c8f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.829231] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5145999, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.86141} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.829938] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.831877] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74406ddf-45e6-4eb0-bba7-c42facbe241c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.856840] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.857248] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146003, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.861393] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7e2b6c20-fdc0-4c5b-90a6-5bdf0648ea4d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.882024] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 784.882024] env[66641]: value = "task-5146004" [ 784.882024] env[66641]: _type = "Task" [ 784.882024] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.891358] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 784.891738] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 784.903875] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Reconfiguring VM instance instance-00000027 to attach disk [datastore2] 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db/34a98372-2ab7-4b21-8a0e-2fc3b91ef4db.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.910601] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c3686296-ef14-45d7-b682-31eb5541b506 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.932088] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 784.932088] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 784.932088] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleting the datastore file [datastore1] bc13bbe2-6e02-4d98-9e50-94f772d89ac0 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 784.932334] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 784.932334] env[66641]: value = "task-5146006" [ 784.932334] env[66641]: _type = "Task" [ 784.932334] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.935422] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-178c8c12-c5ee-493a-b79b-697f687943ef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.945824] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146004, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084395} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.948345] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 784.948840] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 784.948840] env[66641]: value = "task-5146007" [ 784.948840] env[66641]: _type = "Task" [ 784.948840] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.950599] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c2f96b-7d07-46de-a8e0-b6d24e9bb966 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.966261] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for the task: (returnval){ [ 784.966261] env[66641]: value = "task-5146008" [ 784.966261] env[66641]: _type = "Task" [ 784.966261] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.966997] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146006, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.994661] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Reconfiguring VM instance instance-00000025 to attach disk [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 784.998256] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146007, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.998256] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-08cc14f1-cd2d-457c-8539-4b90d27c8931 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.018434] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5146008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.027683] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 785.027683] env[66641]: value = "task-5146009" [ 785.027683] env[66641]: _type = "Task" [ 785.027683] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.038418] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146009, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.291768] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.329s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.292509] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 785.296364] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.115s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.296627] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.300262] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.322s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 785.303122] env[66641]: INFO nova.compute.claims [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 785.309026] env[66641]: DEBUG nova.compute.manager [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 785.309854] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.312491] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81f9098d-750d-45d5-ba8f-0eb12197e686 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.329627] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 785.336145] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9465482-bd21-4461-83cd-59dbac2e6791 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.340768] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146003, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.343096] env[66641]: INFO nova.scheduler.client.report [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Deleted allocations for instance 25c7bd59-ec24-4d30-840b-3c4549dbf669 [ 785.354293] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 785.354293] env[66641]: value = "task-5146010" [ 785.354293] env[66641]: _type = "Task" [ 785.354293] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.370119] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5146010, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.405683] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 785.451803] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146006, 'name': ReconfigVM_Task, 'duration_secs': 0.525262} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.453452] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.453452] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 50 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 785.468783] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146007, 'name': ReconfigVM_Task, 'duration_secs': 0.477028} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.472467] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Reconfigured VM instance instance-00000027 to attach disk [datastore2] 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db/34a98372-2ab7-4b21-8a0e-2fc3b91ef4db.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.473335] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e89cab4-fa39-4bc7-af3e-5b492210d05a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.482532] env[66641]: DEBUG oslo_vmware.api [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Task: {'id': task-5146008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208466} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.484271] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 785.484618] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 785.484725] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.485155] env[66641]: INFO nova.compute.manager [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Took 1.89 seconds to destroy the instance on the hypervisor. [ 785.485155] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 785.485401] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 785.485401] env[66641]: value = "task-5146011" [ 785.485401] env[66641]: _type = "Task" [ 785.485401] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.485594] env[66641]: DEBUG nova.compute.manager [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 785.485688] env[66641]: DEBUG nova.network.neutron [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 785.486248] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.486505] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.512848] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146011, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.536682] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146009, 'name': ReconfigVM_Task, 'duration_secs': 0.464861} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.539303] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.539564] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.549525] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Reconfigured VM instance instance-00000025 to attach disk [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10/081457ae-e152-410c-bca7-4d43b95eee10.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 785.550767] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4a27f27-0c80-472b-b2e5-cb07e32a2985 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.558775] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 785.558775] env[66641]: value = "task-5146012" [ 785.558775] env[66641]: _type = "Task" [ 785.558775] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.572785] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146012, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.769176] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.769597] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.812388] env[66641]: DEBUG nova.compute.utils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 785.817787] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 785.829670] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146003, 'name': CreateSnapshot_Task, 'duration_secs': 0.885686} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.831493] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 785.832335] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06cbe8e7-8991-417f-9478-ea0c99ea7167 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.856865] env[66641]: DEBUG oslo_concurrency.lockutils [None req-9425b3b8-7759-42ab-95ff-97239b3c5fa9 tempest-SecurityGroupsTestJSON-2114626322 tempest-SecurityGroupsTestJSON-2114626322-project-member] Lock "25c7bd59-ec24-4d30-840b-3c4549dbf669" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.129s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 785.867056] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5146010, 'name': PowerOffVM_Task, 'duration_secs': 0.313526} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.867255] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 785.867491] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 785.867762] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46d967c7-1989-4c22-9a0d-ea14ac5abbf7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.932247] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 785.949716] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 785.949843] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 785.950141] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleting the datastore file [datastore1] 05882781-78be-4568-95f4-2fccc4cf4dfe {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.950528] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6089d5c-1137-4c21-9794-0658102f523e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.955166] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 785.955552] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 785.966508] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91993327-347f-4fd6-a675-fe9bd91c7d62 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.969228] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for the task: (returnval){ [ 785.969228] env[66641]: value = "task-5146014" [ 785.969228] env[66641]: _type = "Task" [ 785.969228] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.988184] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c8f9d94-dc11-4b79-912e-c991408e8bdc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.993316] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5146014, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.012239] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 67 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 786.026958] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146011, 'name': Rename_Task, 'duration_secs': 0.264542} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.027272] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 786.027528] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b95ec6d6-105d-463c-a9d5-cf8f2141dbe6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.036146] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 786.036146] env[66641]: value = "task-5146015" [ 786.036146] env[66641]: _type = "Task" [ 786.036146] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.048413] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146015, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.068711] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146012, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.089778] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.090170] env[66641]: WARNING openstack [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.131167] env[66641]: DEBUG nova.compute.manager [req-84eff621-1084-4874-8f06-c4d79fab6ea5 req-eb8e11a9-86ec-4c13-a7c0-597d1109121e service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Received event network-vif-deleted-653c0e4c-f1b9-44f8-abcc-6d580fd70565 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 786.131167] env[66641]: INFO nova.compute.manager [req-84eff621-1084-4874-8f06-c4d79fab6ea5 req-eb8e11a9-86ec-4c13-a7c0-597d1109121e service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Neutron deleted interface 653c0e4c-f1b9-44f8-abcc-6d580fd70565; detaching it from the instance and deleting it from the info cache [ 786.131167] env[66641]: DEBUG nova.network.neutron [req-84eff621-1084-4874-8f06-c4d79fab6ea5 req-eb8e11a9-86ec-4c13-a7c0-597d1109121e service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 786.259278] env[66641]: DEBUG nova.network.neutron [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [{"id": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "address": "fa:16:3e:e9:31:cc", "network": {"id": "3df6a80a-4b35-4871-9321-b0e913258005", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1939361732-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8449c95dd0f74492929e4d1d492ac9d2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a8b99a46-3e7f-4ef1-9e45-58e6cd17f210", "external-id": "nsx-vlan-transportzone-704", "segmentation_id": 704, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2f5f78fa-cb", "ovs_interfaceid": "2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 786.319515] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 786.331480] env[66641]: DEBUG nova.network.neutron [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 786.370699] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 786.372018] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-778715a8-8bba-4ea4-8a2f-0bacb4f16239 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.382746] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 786.382746] env[66641]: value = "task-5146016" [ 786.382746] env[66641]: _type = "Task" [ 786.382746] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.395568] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.479894] env[66641]: DEBUG oslo_vmware.api [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Task: {'id': task-5146014, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.197063} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.480186] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.480372] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 786.480548] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 786.480715] env[66641]: INFO nova.compute.manager [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Took 1.17 seconds to destroy the instance on the hypervisor. [ 786.480963] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 786.481309] env[66641]: DEBUG nova.compute.manager [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 786.481416] env[66641]: DEBUG nova.network.neutron [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 786.482072] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.482183] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.523730] env[66641]: WARNING neutronclient.v2_0.client [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 786.548153] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146015, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.559909] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 786.560226] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 786.579947] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146012, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.590094] env[66641]: DEBUG nova.network.neutron [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Port a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 binding to destination host cpu-1 is already ACTIVE {{(pid=66641) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 786.633996] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e737a5f-2498-462b-8078-bb3b47a56778 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.649162] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f3c359a-9f85-4405-a8c8-649b56b33367 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.690242] env[66641]: DEBUG nova.compute.manager [req-84eff621-1084-4874-8f06-c4d79fab6ea5 req-eb8e11a9-86ec-4c13-a7c0-597d1109121e service nova] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Detach interface failed, port_id=653c0e4c-f1b9-44f8-abcc-6d580fd70565, reason: Instance bc13bbe2-6e02-4d98-9e50-94f772d89ac0 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 786.692533] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb2d35fc-07d9-4cb6-875c-c13f1ef0bc20 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.702013] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa46ef6f-0737-40dc-9f8c-39e589f50092 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.741075] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64e039f7-028d-48b8-bf0a-2c259d7043e9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.753141] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834be08a-3a17-4446-8662-779e8eb07fe1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.771184] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Releasing lock "refresh_cache-047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 786.771580] env[66641]: DEBUG nova.objects.instance [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lazy-loading 'migration_context' on Instance uuid 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 786.773639] env[66641]: DEBUG nova.compute.provider_tree [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.837295] env[66641]: INFO nova.compute.manager [-] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Took 1.35 seconds to deallocate network for instance. [ 786.897985] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.050468] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146015, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.078972] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146012, 'name': Rename_Task, 'duration_secs': 1.351747} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.078972] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 787.079839] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0134b67a-140d-4462-83c5-a8d2bd33d7be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.086885] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 787.086885] env[66641]: value = "task-5146017" [ 787.086885] env[66641]: _type = "Task" [ 787.086885] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.109638] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146017, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.165276] env[66641]: DEBUG nova.compute.manager [req-2be3b84c-f5ec-4e72-8478-69790fac02c4 req-8a3aa2e3-3049-479a-bec1-f9ad57adb8f4 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Received event network-vif-deleted-923d6e74-642d-491f-bf2f-c3768d2d6e85 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 787.165519] env[66641]: INFO nova.compute.manager [req-2be3b84c-f5ec-4e72-8478-69790fac02c4 req-8a3aa2e3-3049-479a-bec1-f9ad57adb8f4 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Neutron deleted interface 923d6e74-642d-491f-bf2f-c3768d2d6e85; detaching it from the instance and deleting it from the info cache [ 787.165884] env[66641]: DEBUG nova.network.neutron [req-2be3b84c-f5ec-4e72-8478-69790fac02c4 req-8a3aa2e3-3049-479a-bec1-f9ad57adb8f4 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.277541] env[66641]: DEBUG nova.scheduler.client.report [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 787.283505] env[66641]: DEBUG nova.objects.base [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Object Instance<047a5c42-3930-4e6a-b3a5-5dbf55d44a4f> lazy-loaded attributes: info_cache,migration_context {{(pid=66641) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 787.286441] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ec4590b-0d5f-422e-8ef3-8c4b631b4f66 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.309858] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a5e4f49-60cb-4270-8084-c8fe182cb182 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.319018] env[66641]: DEBUG oslo_vmware.api [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 787.319018] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e8db2c-69db-3644-8195-76dbe646585e" [ 787.319018] env[66641]: _type = "Task" [ 787.319018] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.329483] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 787.339052] env[66641]: DEBUG oslo_vmware.api [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e8db2c-69db-3644-8195-76dbe646585e, 'name': SearchDatastore_Task} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.339052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.347841] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.366562] env[66641]: DEBUG nova.network.neutron [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 787.381365] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 787.382364] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 787.382364] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 787.382364] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 787.382364] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 787.382364] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 787.382655] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.382655] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 787.382861] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 787.383012] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 787.383121] env[66641]: DEBUG nova.virt.hardware [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 787.384120] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7fe12e-4561-462a-a04e-fad35e823401 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.404247] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.405108] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5532ffaf-b428-40ed-8770-fdb361bfed34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.424921] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 787.431329] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Creating folder: Project (48d0743535fe4211990f170bcd69c702). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.431593] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8092c362-6243-4348-8567-95695c0cbfb0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.445169] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Created folder: Project (48d0743535fe4211990f170bcd69c702) in parent group-v1000566. [ 787.445356] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Creating folder: Instances. Parent ref: group-v1000683. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 787.445656] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cf5b019d-a0e2-48e2-aead-33ae11a7ba6d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.457996] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Created folder: Instances in parent group-v1000683. [ 787.458352] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 787.458599] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 787.458835] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a9ad01f3-ee18-4d8b-bfb3-c74432a0de00 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.480551] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 787.480551] env[66641]: value = "task-5146020" [ 787.480551] env[66641]: _type = "Task" [ 787.480551] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.492450] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146020, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.549619] env[66641]: DEBUG oslo_vmware.api [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146015, 'name': PowerOnVM_Task, 'duration_secs': 1.355338} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.549926] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 787.550266] env[66641]: INFO nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Took 10.34 seconds to spawn the instance on the hypervisor. [ 787.550381] env[66641]: DEBUG nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 787.552059] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c907a27-e225-4029-8edc-097d52f8a124 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.597729] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146017, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.634235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.634235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.634235] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.669369] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71cf3689-f506-47b5-8ae5-d698ed8a5209 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.681759] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f90d5cf-36ab-43b0-a6bc-c16ac1f17ec2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.719817] env[66641]: DEBUG nova.compute.manager [req-2be3b84c-f5ec-4e72-8478-69790fac02c4 req-8a3aa2e3-3049-479a-bec1-f9ad57adb8f4 service nova] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Detach interface failed, port_id=923d6e74-642d-491f-bf2f-c3768d2d6e85, reason: Instance 05882781-78be-4568-95f4-2fccc4cf4dfe could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 787.742161] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "dad5ed23-71a1-4b55-856f-2484f8e62708" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.742504] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.742768] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 787.742978] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.743208] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.745846] env[66641]: INFO nova.compute.manager [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Terminating instance [ 787.786991] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.787257] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 787.790416] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.341s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.790615] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 787.792885] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.351s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 787.793125] env[66641]: DEBUG nova.objects.instance [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 787.830544] env[66641]: INFO nova.scheduler.client.report [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Deleted allocations for instance 0f6055b9-f5b4-48ba-9589-0af212808be7 [ 787.870769] env[66641]: INFO nova.compute.manager [-] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Took 1.39 seconds to deallocate network for instance. [ 787.901770] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.990117] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146020, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.073352] env[66641]: INFO nova.compute.manager [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Took 28.86 seconds to build instance. [ 788.101872] env[66641]: DEBUG oslo_vmware.api [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146017, 'name': PowerOnVM_Task, 'duration_secs': 0.619955} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.102724] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 788.102724] env[66641]: DEBUG nova.compute.manager [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 788.103621] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-968e2c63-41de-421a-b4e4-97b1fa229695 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.249781] env[66641]: DEBUG nova.compute.manager [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 788.250196] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 788.251582] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a8ce36-64be-45af-bb6f-575d746979e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.263135] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 788.263482] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fd58c3d5-b3ed-44de-915b-0d261f9d9cc4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.271513] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 788.271513] env[66641]: value = "task-5146021" [ 788.271513] env[66641]: _type = "Task" [ 788.271513] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.281438] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5146021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.305875] env[66641]: DEBUG nova.compute.utils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 788.311906] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 788.312506] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 788.312680] env[66641]: WARNING neutronclient.v2_0.client [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.313038] env[66641]: WARNING neutronclient.v2_0.client [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 788.313833] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.313982] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.340582] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b5cd1ecd-84fa-4611-a9b1-1f35a4b02bfd tempest-ImagesOneServerNegativeTestJSON-204585410 tempest-ImagesOneServerNegativeTestJSON-204585410-project-member] Lock "0f6055b9-f5b4-48ba-9589-0af212808be7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.686s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.376699] env[66641]: DEBUG nova.policy [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '942a73a2c987416d8cf6c66e89e3ab34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '17276c1118a6476b81390021f8fe9ed2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 788.383254] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.399419] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.493844] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146020, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.576672] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c3f93f06-ed2a-4d89-85b8-e08337f8a76f tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.382s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.626750] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 788.642686] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 788.643560] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 788.722224] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Successfully created port: 5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 788.782657] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5146021, 'name': PowerOffVM_Task, 'duration_secs': 0.307431} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.784786] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 788.784786] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 788.784786] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59d72dc7-7af5-4eb4-ad7e-26bb3af519aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.812291] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 788.817232] env[66641]: DEBUG oslo_concurrency.lockutils [None req-37fc31af-ee9d-49d2-ad4c-cf1f9d00ebbf tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.024s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 788.820310] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 5.847s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 788.853178] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 788.853441] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 788.853692] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Deleting the datastore file [datastore1] dad5ed23-71a1-4b55-856f-2484f8e62708 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 788.854390] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-723a78c6-3624-406a-93ec-4d2cfa60c74a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.861784] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for the task: (returnval){ [ 788.861784] env[66641]: value = "task-5146023" [ 788.861784] env[66641]: _type = "Task" [ 788.861784] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 788.873171] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5146023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.900805] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.940886] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.940886] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.940886] env[66641]: DEBUG nova.network.neutron [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 788.996386] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146020, 'name': CreateVM_Task, 'duration_secs': 1.256905} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 788.996641] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 788.997236] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.997469] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 788.997928] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 788.998423] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25389e61-264b-4ea6-952f-b6f7d505093e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.005942] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 789.005942] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52246597-ab55-eae2-aa06-42747a7157f4" [ 789.005942] env[66641]: _type = "Task" [ 789.005942] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.019717] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52246597-ab55-eae2-aa06-42747a7157f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.333388] env[66641]: INFO nova.compute.claims [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.373653] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5146023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.402475] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.447022] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.447022] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.517743] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52246597-ab55-eae2-aa06-42747a7157f4, 'name': SearchDatastore_Task, 'duration_secs': 0.014239} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.518258] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 789.518570] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 789.518870] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.519086] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 789.519330] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 789.519666] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-51ba2c98-0e54-4b36-94c6-9285b3f157db {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.531055] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 789.531271] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 789.532109] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c887c91-71f9-446a-b4f2-8f9ff7499ad3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.542034] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 789.542034] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5265ba5a-1962-6e6b-e39f-db76b7151d37" [ 789.542034] env[66641]: _type = "Task" [ 789.542034] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.549743] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5265ba5a-1962-6e6b-e39f-db76b7151d37, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.840822] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 789.844878] env[66641]: INFO nova.compute.resource_tracker [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating resource usage from migration e82278e1-5e69-4b99-8727-b88245abb346 [ 789.878483] env[66641]: DEBUG oslo_vmware.api [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Task: {'id': task-5146023, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.675268} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.881093] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 789.881093] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 789.881299] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 789.881443] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 789.881549] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 789.881680] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 789.881880] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 789.882042] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 789.882208] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 789.882422] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 789.882574] env[66641]: DEBUG nova.virt.hardware [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 789.883027] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 789.883160] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 789.883222] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.883428] env[66641]: INFO nova.compute.manager [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Took 1.63 seconds to destroy the instance on the hypervisor. [ 789.883657] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 789.884588] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac768dfb-4230-4257-addc-323ded9a4c21 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.892315] env[66641]: DEBUG nova.compute.manager [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 789.892461] env[66641]: DEBUG nova.network.neutron [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 789.893300] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 789.894113] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 789.913414] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebbe410c-1fec-4133-a18d-32e25c534c2c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.921186] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.053015] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5265ba5a-1962-6e6b-e39f-db76b7151d37, 'name': SearchDatastore_Task, 'duration_secs': 0.027989} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.056974] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06121f96-f9b4-4a31-bebc-602bd8b2474c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.064129] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 790.064129] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f2083f-27a8-1140-d0cc-9f99f566a288" [ 790.064129] env[66641]: _type = "Task" [ 790.064129] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.077271] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f2083f-27a8-1140-d0cc-9f99f566a288, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.192902] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cbcb70-26b9-43b5-826b-1fdc5254d41b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.201406] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ab290b-b5ec-4f34-b6a1-66bd4ea4d380 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.237716] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101aa234-8087-4a4e-9af9-c9869eee2dea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.247876] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e292b3a-b64a-4908-8d29-8bb19fe8b699 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.265929] env[66641]: DEBUG nova.compute.provider_tree [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.411962] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.482195] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.482799] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.575395] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Successfully updated port: 5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 790.591159] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f2083f-27a8-1140-d0cc-9f99f566a288, 'name': SearchDatastore_Task, 'duration_secs': 0.01361} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.594329] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 790.594329] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 207483d3-803c-495b-9b93-6f986f3ca56e/207483d3-803c-495b-9b93-6f986f3ca56e.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 790.594329] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-76bed157-1ef3-4998-ade8-75e8ceb098b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.608309] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 790.608309] env[66641]: value = "task-5146024" [ 790.608309] env[66641]: _type = "Task" [ 790.608309] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.620120] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.769988] env[66641]: DEBUG nova.scheduler.client.report [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 790.839630] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.839983] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 790.913657] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.918747] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 790.919299] env[66641]: WARNING openstack [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.045803] env[66641]: DEBUG nova.network.neutron [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.080702] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.080702] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquired lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.080702] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 791.123783] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146024, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.277820] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.455s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.277820] env[66641]: INFO nova.compute.manager [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Migrating [ 791.284124] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.106s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.285994] env[66641]: INFO nova.compute.claims [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.327251] env[66641]: DEBUG nova.network.neutron [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 791.353107] env[66641]: DEBUG nova.compute.manager [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Received event network-vif-plugged-5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 791.353334] env[66641]: DEBUG oslo_concurrency.lockutils [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] Acquiring lock "40d7571f-09e2-463a-a449-36c621045819-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.353725] env[66641]: DEBUG oslo_concurrency.lockutils [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] Lock "40d7571f-09e2-463a-a449-36c621045819-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.353987] env[66641]: DEBUG oslo_concurrency.lockutils [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] Lock "40d7571f-09e2-463a-a449-36c621045819-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.354258] env[66641]: DEBUG nova.compute.manager [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] No waiting events found dispatching network-vif-plugged-5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 791.354456] env[66641]: WARNING nova.compute.manager [req-3fd5e8d6-c929-4913-9619-72aaef332c20 req-c98ac878-badb-4d4d-9b87-08d14b555fa3 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Received unexpected event network-vif-plugged-5f780526-bca9-4384-a9fc-e10c3d80fb62 for instance with vm_state building and task_state spawning. [ 791.414372] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146016, 'name': CloneVM_Task, 'duration_secs': 4.590831} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.415754] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Created linked-clone VM from snapshot [ 791.418593] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a705463-7d23-4cbe-a125-6c9722e6c0be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.428859] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Uploading image 2875e846-84c5-4d52-b64d-3ed1734800e4 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 791.458775] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 791.458775] env[66641]: value = "vm-1000682" [ 791.458775] env[66641]: _type = "VirtualMachine" [ 791.458775] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 791.459425] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-5ef9d9cd-a22c-47b5-b920-cbfbcaf43901 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.470652] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease: (returnval){ [ 791.470652] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ffb7b6-e8dd-529d-0e0b-6f9e96608e9e" [ 791.470652] env[66641]: _type = "HttpNfcLease" [ 791.470652] env[66641]: } obtained for exporting VM: (result){ [ 791.470652] env[66641]: value = "vm-1000682" [ 791.470652] env[66641]: _type = "VirtualMachine" [ 791.470652] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 791.471152] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the lease: (returnval){ [ 791.471152] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ffb7b6-e8dd-529d-0e0b-6f9e96608e9e" [ 791.471152] env[66641]: _type = "HttpNfcLease" [ 791.471152] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 791.479214] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.479214] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ffb7b6-e8dd-529d-0e0b-6f9e96608e9e" [ 791.479214] env[66641]: _type = "HttpNfcLease" [ 791.479214] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 791.551361] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 791.583156] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.583770] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.592706] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 791.626599] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146024, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589522} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.634032] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 207483d3-803c-495b-9b93-6f986f3ca56e/207483d3-803c-495b-9b93-6f986f3ca56e.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 791.634032] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 791.634032] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66e9624a-eb93-457a-afcc-28d3b48fcd71 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.643214] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 791.643214] env[66641]: value = "task-5146026" [ 791.643214] env[66641]: _type = "Task" [ 791.643214] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.657305] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146026, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.750575] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.751210] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.806256] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.807045] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 791.807350] env[66641]: DEBUG nova.network.neutron [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 791.828812] env[66641]: INFO nova.compute.manager [-] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Took 1.94 seconds to deallocate network for instance. [ 791.851408] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 791.851813] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 791.925375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "081457ae-e152-410c-bca7-4d43b95eee10" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.925375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.925375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "081457ae-e152-410c-bca7-4d43b95eee10-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 791.925375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 791.925861] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 791.927054] env[66641]: INFO nova.compute.manager [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Terminating instance [ 791.981639] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 791.981639] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ffb7b6-e8dd-529d-0e0b-6f9e96608e9e" [ 791.981639] env[66641]: _type = "HttpNfcLease" [ 791.981639] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 791.981797] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 791.981797] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ffb7b6-e8dd-529d-0e0b-6f9e96608e9e" [ 791.981797] env[66641]: _type = "HttpNfcLease" [ 791.981797] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 791.982789] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b0d62e-aae3-4867-a4e1-6db480a31dee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.992694] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 791.992898] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 792.057842] env[66641]: DEBUG nova.network.neutron [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Updating instance_info_cache with network_info: [{"id": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "address": "fa:16:3e:0c:fe:19", "network": {"id": "9b5685d7-acf0-4ec7-8593-02d91ea193fd", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1735113740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17276c1118a6476b81390021f8fe9ed2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f780526-bc", "ovs_interfaceid": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.087534] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcb1e20-21c6-43a3-a256-5f72f08818d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.112336] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a2e55b-e38e-4924-9da6-c341233f5f23 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.122498] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.122790] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.122995] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.123268] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.123383] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 792.125265] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 83 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 792.129769] env[66641]: INFO nova.compute.manager [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Terminating instance [ 792.159200] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146026, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124925} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.159506] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 792.160377] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fd666e-68dc-4fb8-889f-de67c3f36010 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.183374] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 207483d3-803c-495b-9b93-6f986f3ca56e/207483d3-803c-495b-9b93-6f986f3ca56e.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 792.183933] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8c38831-1226-4d6f-89e9-9835265d1f37 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.208054] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 792.208054] env[66641]: value = "task-5146027" [ 792.208054] env[66641]: _type = "Task" [ 792.208054] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.219756] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146027, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.224950] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c12c727-ad5c-470e-8cd0-707d65a578bd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.318758] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.318758] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.339025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.432028] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "refresh_cache-081457ae-e152-410c-bca7-4d43b95eee10" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.433648] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquired lock "refresh_cache-081457ae-e152-410c-bca7-4d43b95eee10" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 792.433933] env[66641]: DEBUG nova.network.neutron [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 792.466068] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.466068] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.561405] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Releasing lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 792.561719] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Instance network_info: |[{"id": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "address": "fa:16:3e:0c:fe:19", "network": {"id": "9b5685d7-acf0-4ec7-8593-02d91ea193fd", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1735113740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17276c1118a6476b81390021f8fe9ed2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f780526-bc", "ovs_interfaceid": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 792.563171] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0c:fe:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8a34aa30-95be-4b18-98ca-1f2d81f7e9e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f780526-bca9-4384-a9fc-e10c3d80fb62', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 792.570183] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Creating folder: Project (17276c1118a6476b81390021f8fe9ed2). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.574461] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.574855] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.582592] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-48812926-3c56-4223-95c6-efa8f5b8936b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.601796] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Created folder: Project (17276c1118a6476b81390021f8fe9ed2) in parent group-v1000566. [ 792.602059] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Creating folder: Instances. Parent ref: group-v1000686. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 792.604967] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea645d14-6b7a-475a-b1ae-aa338f7153be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.619100] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Created folder: Instances in parent group-v1000686. [ 792.619766] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 792.619766] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 792.620257] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c28e088b-31de-4220-80b4-b7eb32142982 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.647689] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 792.648471] env[66641]: DEBUG nova.compute.manager [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 792.648715] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.649306] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-52b0adfd-9b84-4584-9e3f-4a776fefea33 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.652950] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eedfce95-0831-48ab-9d22-bffad1540f8c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.663523] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 792.671894] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c037e87b-fa6c-4a76-b5bd-0795e8f7be3f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.673812] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 792.673812] env[66641]: value = "task-5146031" [ 792.673812] env[66641]: _type = "Task" [ 792.673812] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.674177] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 792.674177] env[66641]: value = "task-5146030" [ 792.674177] env[66641]: _type = "Task" [ 792.674177] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.689936] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 792.689936] env[66641]: value = "task-5146032" [ 792.689936] env[66641]: _type = "Task" [ 792.689936] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.706506] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146030, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.706789] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146031, 'name': CreateVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.718710] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146032, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.726332] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146027, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.783521] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f91a98-2a6b-4f3d-96fa-6b27f6f45c6c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.793773] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae069f1-2aed-43fc-8dbe-e481f395c598 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.835401] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd4bd9b-9317-4ea8-96e7-f40c1bc90430 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.844308] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 792.845102] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 792.854320] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e85719-acfd-4cbb-af88-2a39ec29f11e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.874793] env[66641]: DEBUG nova.compute.provider_tree [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.879468] env[66641]: DEBUG nova.network.neutron [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 792.937619] env[66641]: WARNING openstack [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 792.938216] env[66641]: WARNING openstack [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 792.945960] env[66641]: DEBUG nova.network.neutron [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 793.194184] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146031, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.204668] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146030, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.213947] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146032, 'name': PowerOffVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.229941] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146027, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.292953] env[66641]: DEBUG nova.network.neutron [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 793.348583] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 793.383407] env[66641]: DEBUG nova.scheduler.client.report [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 793.388008] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.529715] env[66641]: DEBUG nova.compute.manager [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Received event network-changed-5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 793.530959] env[66641]: DEBUG nova.compute.manager [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Refreshing instance network info cache due to event network-changed-5f780526-bca9-4384-a9fc-e10c3d80fb62. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 793.530959] env[66641]: DEBUG oslo_concurrency.lockutils [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Acquiring lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.530959] env[66641]: DEBUG oslo_concurrency.lockutils [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Acquired lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.530959] env[66641]: DEBUG nova.network.neutron [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Refreshing network info cache for port 5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 793.697895] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146031, 'name': CreateVM_Task, 'duration_secs': 0.573604} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.698281] env[66641]: DEBUG oslo_vmware.api [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146030, 'name': PowerOnVM_Task, 'duration_secs': 0.55894} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.703213] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 793.703643] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 793.703930] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d319aa53-a15d-4cbc-bd4a-761a8665330a tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance '58fefaa4-0b17-408f-9329-78f8b5cf3fa7' progress to 100 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 793.709556] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 793.711237] env[66641]: WARNING openstack [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 793.719506] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.723024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.723024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 793.723024] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f230eeb-5957-4c34-b9ff-0b5f2e73ffe0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.734684] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146032, 'name': PowerOffVM_Task, 'duration_secs': 0.684853} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.735965] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 793.736337] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 793.737571] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-286ac185-74f6-4516-a9f4-0138e95c4499 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.744973] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 793.744973] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5260d71b-24ee-3db1-c98f-6eb6706df4e8" [ 793.744973] env[66641]: _type = "Task" [ 793.744973] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.745792] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146027, 'name': ReconfigVM_Task, 'duration_secs': 1.279913} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.746744] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 207483d3-803c-495b-9b93-6f986f3ca56e/207483d3-803c-495b-9b93-6f986f3ca56e.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 793.750907] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f03c625-0dad-43a4-b723-38fb97423fe6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.759917] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5260d71b-24ee-3db1-c98f-6eb6706df4e8, 'name': SearchDatastore_Task, 'duration_secs': 0.013562} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.761904] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.762438] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 793.762786] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.763112] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 793.763488] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 793.763846] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 793.763846] env[66641]: value = "task-5146034" [ 793.763846] env[66641]: _type = "Task" [ 793.763846] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.764186] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0934059f-ae26-4d22-b900-65322d303782 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.777561] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146034, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.782287] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 793.782287] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 793.783700] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1273b5d-1a33-4140-a9fb-94db1b45cd7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.790133] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 793.790133] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522743d7-12ee-ceac-34f9-b1b249d3b25d" [ 793.790133] env[66641]: _type = "Task" [ 793.790133] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.795660] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Releasing lock "refresh_cache-081457ae-e152-410c-bca7-4d43b95eee10" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 793.796369] env[66641]: DEBUG nova.compute.manager [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 793.796645] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.802658] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a884980-4998-44cf-8f7e-34841cab7098 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.805946] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522743d7-12ee-ceac-34f9-b1b249d3b25d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.812587] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 793.812957] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab69a81a-4a34-42c7-a4b3-92c937994dd9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.821216] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 793.821216] env[66641]: value = "task-5146035" [ 793.821216] env[66641]: _type = "Task" [ 793.821216] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.827250] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 793.827636] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 793.827994] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Deleting the datastore file [datastore2] 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 793.829141] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09ce6de5-a0ed-45f1-9413-6f397b237eb9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.837968] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146035, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.840972] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for the task: (returnval){ [ 793.840972] env[66641]: value = "task-5146036" [ 793.840972] env[66641]: _type = "Task" [ 793.840972] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.852812] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146036, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.878495] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 793.896248] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.612s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 793.897101] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 793.902809] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.146s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 793.902809] env[66641]: DEBUG nova.objects.instance [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lazy-loading 'resources' on Instance uuid b9d032da-031e-42e0-86e2-95254c1ceac1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 794.036449] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.037457] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.229341] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.229842] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.279056] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146034, 'name': Rename_Task, 'duration_secs': 0.241737} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.279367] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 794.279648] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-56c6b46b-6ab5-4e95-a31d-9632d6adf395 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.288269] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 794.288269] env[66641]: value = "task-5146037" [ 794.288269] env[66641]: _type = "Task" [ 794.288269] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.298898] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146037, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.309693] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]522743d7-12ee-ceac-34f9-b1b249d3b25d, 'name': SearchDatastore_Task, 'duration_secs': 0.021532} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.310817] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-736c16ed-b8e4-435c-8472-9708d2d8423f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.317894] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 794.317894] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52af70cc-39fa-5e67-77ef-91c196dafa0b" [ 794.317894] env[66641]: _type = "Task" [ 794.317894] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.340734] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146035, 'name': PowerOffVM_Task, 'duration_secs': 0.185274} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.340734] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52af70cc-39fa-5e67-77ef-91c196dafa0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.340981] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 794.340981] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 794.341966] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e10da8ac-528b-4534-a991-302985d2159f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.350936] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.351307] env[66641]: WARNING openstack [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.366659] env[66641]: DEBUG oslo_vmware.api [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Task: {'id': task-5146036, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.294297} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.367045] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.367251] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.367423] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.367587] env[66641]: INFO nova.compute.manager [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Took 1.72 seconds to destroy the instance on the hypervisor. [ 794.367839] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 794.368056] env[66641]: DEBUG nova.compute.manager [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 794.368151] env[66641]: DEBUG nova.network.neutron [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 794.368660] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.368916] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.377338] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 794.377544] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 794.377718] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Deleting the datastore file [datastore2] 081457ae-e152-410c-bca7-4d43b95eee10 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.378363] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a787625-632a-402a-a34e-054f43f4ea56 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.385988] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for the task: (returnval){ [ 794.385988] env[66641]: value = "task-5146039" [ 794.385988] env[66641]: _type = "Task" [ 794.385988] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.395407] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146039, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.406713] env[66641]: DEBUG nova.compute.utils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 794.411524] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 794.412055] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 794.412456] env[66641]: WARNING neutronclient.v2_0.client [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.412720] env[66641]: WARNING neutronclient.v2_0.client [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 794.414090] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.414090] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.452442] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.454041] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.490587] env[66641]: DEBUG nova.network.neutron [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Updated VIF entry in instance network info cache for port 5f780526-bca9-4384-a9fc-e10c3d80fb62. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 794.491870] env[66641]: DEBUG nova.network.neutron [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Updating instance_info_cache with network_info: [{"id": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "address": "fa:16:3e:0c:fe:19", "network": {"id": "9b5685d7-acf0-4ec7-8593-02d91ea193fd", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-1735113740-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "17276c1118a6476b81390021f8fe9ed2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8a34aa30-95be-4b18-98ca-1f2d81f7e9e6", "external-id": "nsx-vlan-transportzone-234", "segmentation_id": 234, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f780526-bc", "ovs_interfaceid": "5f780526-bca9-4384-a9fc-e10c3d80fb62", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 794.552127] env[66641]: DEBUG nova.policy [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '173c0e24fb5c4eea8a52e9a8c7e570e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c48b059b6f7d49a8ba5ecf28125ddaa9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 794.805545] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146037, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.832212] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52af70cc-39fa-5e67-77ef-91c196dafa0b, 'name': SearchDatastore_Task, 'duration_secs': 0.018831} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.832212] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 794.832338] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 40d7571f-09e2-463a-a449-36c621045819/40d7571f-09e2-463a-a449-36c621045819.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 794.832644] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6c9a08ac-f869-48df-85f7-d83918e20d37 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.841781] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 794.841781] env[66641]: value = "task-5146040" [ 794.841781] env[66641]: _type = "Task" [ 794.841781] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.864013] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146040, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.881340] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bd62e20-aa1b-43d9-9147-b55196a35f4b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.893231] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ff67ad6-4a79-4935-b58b-bc62c1051879 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.903801] env[66641]: DEBUG oslo_vmware.api [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Task: {'id': task-5146039, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175787} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.932014] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 794.932014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 794.932014] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.932014] env[66641]: INFO nova.compute.manager [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Took 1.13 seconds to destroy the instance on the hypervisor. [ 794.932014] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 794.940628] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 794.940628] env[66641]: DEBUG nova.compute.manager [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 794.940628] env[66641]: DEBUG nova.network.neutron [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 794.940628] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 794.940628] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 794.947120] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d70e90ae-8067-4951-b9b1-a13d2594da63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.950833] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff43df6-2f48-4d23-a811-f39e7b7fbf34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.972799] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 0 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 794.983240] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc97b79-98a3-482e-b552-240870e47362 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.999622] env[66641]: DEBUG oslo_concurrency.lockutils [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] Releasing lock "refresh_cache-40d7571f-09e2-463a-a449-36c621045819" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 795.000314] env[66641]: DEBUG nova.compute.manager [req-38b697e1-2614-4421-8b17-352ee5d704a6 req-7e325fc1-0c17-43f3-bad1-588d730730bd service nova] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Received event network-vif-deleted-f6c1b16b-1415-4eaf-9228-e8c95836606c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 795.000843] env[66641]: DEBUG nova.compute.provider_tree [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.295524] env[66641]: DEBUG nova.network.neutron [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.307323] env[66641]: DEBUG oslo_vmware.api [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146037, 'name': PowerOnVM_Task, 'duration_secs': 0.62369} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.307636] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 795.307841] env[66641]: INFO nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Took 7.98 seconds to spawn the instance on the hypervisor. [ 795.308028] env[66641]: DEBUG nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 795.309140] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf2dcc6-0883-425d-ab7f-2778db04ba80 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.361789] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146040, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.398228] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Successfully created port: e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 795.450937] env[66641]: DEBUG nova.network.neutron [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 795.452753] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 795.452753] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 795.482238] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 795.483446] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cb17bdde-8882-40b5-b9f6-3f9d0617a652 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.495828] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 795.495828] env[66641]: value = "task-5146041" [ 795.495828] env[66641]: _type = "Task" [ 795.495828] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.504263] env[66641]: DEBUG nova.scheduler.client.report [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 795.515230] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146041, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.801609] env[66641]: INFO nova.compute.manager [-] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Took 1.43 seconds to deallocate network for instance. [ 795.832618] env[66641]: INFO nova.compute.manager [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Took 27.36 seconds to build instance. [ 795.858826] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146040, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655242} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.862850] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 40d7571f-09e2-463a-a449-36c621045819/40d7571f-09e2-463a-a449-36c621045819.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 795.862850] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 795.862850] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-81c0aab8-1ca9-4ed4-9c01-5adb8440d744 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.872167] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 795.872167] env[66641]: value = "task-5146042" [ 795.872167] env[66641]: _type = "Task" [ 795.872167] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.886047] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146042, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.946171] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 795.959940] env[66641]: DEBUG nova.network.neutron [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 795.978952] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 795.978952] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 795.980895] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 795.980974] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 795.981233] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 795.981435] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 795.981733] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.981918] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 795.982166] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 795.982629] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 795.982908] env[66641]: DEBUG nova.virt.hardware [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 795.985130] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63944ea6-1ca1-42e7-87f5-f319702f1d37 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.996617] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6614ea80-92be-4688-a57a-ab7cc4f69aa2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.010250] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.110s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.024882] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.262s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 796.025123] env[66641]: DEBUG nova.objects.instance [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lazy-loading 'resources' on Instance uuid 61042df0-a727-4aa8-b2ea-bdc40899d0fc {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.029516] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146041, 'name': PowerOffVM_Task, 'duration_secs': 0.234059} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.029516] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 796.029516] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 17 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 796.036311] env[66641]: INFO nova.scheduler.client.report [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Deleted allocations for instance b9d032da-031e-42e0-86e2-95254c1ceac1 [ 796.314492] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.338312] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6d01769a-0571-427f-830a-b598b990e110 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.870s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.383821] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146042, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.150835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.384309] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 796.385167] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda9d5d4-cb21-459a-b457-8bff4ea896b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.410675] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 40d7571f-09e2-463a-a449-36c621045819/40d7571f-09e2-463a-a449-36c621045819.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 796.411023] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05075335-8293-42dc-a338-9a6d087600e8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.431670] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 796.431670] env[66641]: value = "task-5146043" [ 796.431670] env[66641]: _type = "Task" [ 796.431670] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.440823] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146043, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.463655] env[66641]: INFO nova.compute.manager [-] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Took 1.53 seconds to deallocate network for instance. [ 796.536024] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:17Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 796.536024] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 796.536024] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 796.536024] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 796.536336] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 796.536336] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 796.536405] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.536555] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 796.536716] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 796.536869] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 796.537041] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 796.543590] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d54637e1-9b7f-408e-a1f5-cfae37338717 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.561580] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83208cdf-5912-4e3c-b218-fbaaa2f24e5f tempest-ServersV294TestFqdnHostnames-2105284201 tempest-ServersV294TestFqdnHostnames-2105284201-project-member] Lock "b9d032da-031e-42e0-86e2-95254c1ceac1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.979s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 796.570448] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 796.570448] env[66641]: value = "task-5146044" [ 796.570448] env[66641]: _type = "Task" [ 796.570448] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.586190] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146044, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.911299] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6db0320-e95a-4d3e-8de4-7b4c6d607c14 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.920407] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-675ed733-7a0b-4369-88f9-d2a8d05d9c02 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.955934] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59920b81-56b0-4e1d-851d-fbcf5143361e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.967945] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc5e595-b8fe-4fd2-b980-4a7dd023d19c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.972469] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146043, 'name': ReconfigVM_Task, 'duration_secs': 0.532001} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.974029] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 796.975022] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 40d7571f-09e2-463a-a449-36c621045819/40d7571f-09e2-463a-a449-36c621045819.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 796.975415] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c80dbda-5754-4b7c-8223-22dd58d081f6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.986288] env[66641]: DEBUG nova.compute.provider_tree [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.989294] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 796.989294] env[66641]: value = "task-5146045" [ 796.989294] env[66641]: _type = "Task" [ 796.989294] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.999856] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146045, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.081934] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146044, 'name': ReconfigVM_Task, 'duration_secs': 0.245072} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.082320] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 33 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 797.297228] env[66641]: DEBUG nova.compute.manager [req-5abf26d0-2081-4e3a-8fff-c338912faf00 req-355bed02-9581-4081-8cf2-233eea055e44 service nova] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Received event network-vif-deleted-9738572a-f121-4d8c-be4f-771175b83bed {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 797.438792] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Successfully updated port: e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 797.491337] env[66641]: DEBUG nova.scheduler.client.report [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 797.511541] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146045, 'name': Rename_Task, 'duration_secs': 0.310151} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.512901] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 797.513179] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7462ff04-481f-4845-831c-20a72a044b8c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.521620] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 797.521620] env[66641]: value = "task-5146046" [ 797.521620] env[66641]: _type = "Task" [ 797.521620] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.533594] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146046, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.590950] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 797.590950] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 797.590950] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 797.590950] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 797.591204] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 797.591575] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 797.592636] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 797.592982] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 797.595275] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 797.595275] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 797.595275] env[66641]: DEBUG nova.virt.hardware [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 797.600254] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfiguring VM instance instance-00000026 to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 797.600613] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c305c31e-e589-461e-93e3-c40ff884b2d3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.630313] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 797.630313] env[66641]: value = "task-5146047" [ 797.630313] env[66641]: _type = "Task" [ 797.630313] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.646332] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146047, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.650100] env[66641]: DEBUG nova.compute.manager [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Received event network-vif-plugged-e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 797.650100] env[66641]: DEBUG oslo_concurrency.lockutils [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 797.650100] env[66641]: DEBUG oslo_concurrency.lockutils [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 797.650100] env[66641]: DEBUG oslo_concurrency.lockutils [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 797.650100] env[66641]: DEBUG nova.compute.manager [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] No waiting events found dispatching network-vif-plugged-e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 797.650658] env[66641]: WARNING nova.compute.manager [req-64b75173-f374-4ae5-a744-c5655f5c0a3d req-222f3bdf-de31-43d6-8416-caa25f3c785d service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Received unexpected event network-vif-plugged-e873e765-2324-4273-81ba-e543ebc3b867 for instance with vm_state building and task_state spawning. [ 797.942904] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.943520] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquired lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 797.943793] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 798.002981] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.978s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.006211] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.074s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.008895] env[66641]: INFO nova.compute.claims [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.037408] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146046, 'name': PowerOnVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.043602] env[66641]: INFO nova.scheduler.client.report [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Deleted allocations for instance 61042df0-a727-4aa8-b2ea-bdc40899d0fc [ 798.130812] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.131376] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.163512] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146047, 'name': ReconfigVM_Task, 'duration_secs': 0.215503} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.164466] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfigured VM instance instance-00000026 to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 798.165455] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37dec443-aaa7-47f9-b988-98d45ff62b91 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.194447] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfiguring VM instance instance-00000026 to attach disk [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.198265] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.198983] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.205084] env[66641]: WARNING neutronclient.v2_0.client [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 798.208239] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17fb24d3-3e6e-49f5-b5a6-7d3b40089180 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.229054] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 798.229054] env[66641]: value = "task-5146048" [ 798.229054] env[66641]: _type = "Task" [ 798.229054] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.238890] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146048, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.275363] env[66641]: DEBUG nova.network.neutron [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Port a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 binding to destination host cpu-1 is already ACTIVE {{(pid=66641) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 798.276021] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.276021] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 798.276021] env[66641]: DEBUG nova.network.neutron [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 798.447863] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.448298] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.460019] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 798.533458] env[66641]: DEBUG oslo_vmware.api [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146046, 'name': PowerOnVM_Task, 'duration_secs': 0.531816} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.533933] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 798.534202] env[66641]: INFO nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 8.69 seconds to spawn the instance on the hypervisor. [ 798.534425] env[66641]: DEBUG nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 798.535283] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890e223d-81cf-4320-80aa-f58f4c328f04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.554886] env[66641]: DEBUG oslo_concurrency.lockutils [None req-17b8c9eb-f78d-4ba4-8195-2493839cb9c5 tempest-ServersListShow298Test-1868603876 tempest-ServersListShow298Test-1868603876-project-member] Lock "61042df0-a727-4aa8-b2ea-bdc40899d0fc" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.707s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.600534] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.600978] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.666897] env[66641]: DEBUG nova.compute.manager [None req-6963fb5c-e3bf-4de3-8247-8beca3e1d539 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 798.668897] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0107a54-2964-4f11-8db7-ae0ba0287e15 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.712568] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.712568] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.740773] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146048, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.781175] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.781340] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 798.866180] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "207483d3-803c-495b-9b93-6f986f3ca56e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.866458] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.866669] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "207483d3-803c-495b-9b93-6f986f3ca56e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 798.866917] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 798.867024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 798.869735] env[66641]: INFO nova.compute.manager [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Terminating instance [ 798.882406] env[66641]: DEBUG nova.network.neutron [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [{"id": "e873e765-2324-4273-81ba-e543ebc3b867", "address": "fa:16:3e:fb:cd:da", "network": {"id": "3cfa3773-ee60-40a6-af0b-3fa7df1cae1a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1453977923-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c48b059b6f7d49a8ba5ecf28125ddaa9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape873e765-23", "ovs_interfaceid": "e873e765-2324-4273-81ba-e543ebc3b867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 798.972405] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 798.973092] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.055081] env[66641]: INFO nova.compute.manager [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 25.11 seconds to build instance. [ 799.184045] env[66641]: INFO nova.compute.manager [None req-6963fb5c-e3bf-4de3-8247-8beca3e1d539 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] instance snapshotting [ 799.184893] env[66641]: DEBUG nova.objects.instance [None req-6963fb5c-e3bf-4de3-8247-8beca3e1d539 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lazy-loading 'flavor' on Instance uuid 207483d3-803c-495b-9b93-6f986f3ca56e {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 799.242137] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146048, 'name': ReconfigVM_Task, 'duration_secs': 0.638168} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.245439] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Reconfigured VM instance instance-00000026 to attach disk [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5/48f99287-b737-45fa-ad59-9e1425afa3d5.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.245795] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 50 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 799.363402] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c5fe141-529b-4537-a023-d2182cd6ff48 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.374051] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "refresh_cache-207483d3-803c-495b-9b93-6f986f3ca56e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.374425] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquired lock "refresh_cache-207483d3-803c-495b-9b93-6f986f3ca56e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.374608] env[66641]: DEBUG nova.network.neutron [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 799.377085] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d08b6f5-3e64-4eb2-aca9-f0fad3da5e14 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.385056] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Releasing lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 799.385379] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance network_info: |[{"id": "e873e765-2324-4273-81ba-e543ebc3b867", "address": "fa:16:3e:fb:cd:da", "network": {"id": "3cfa3773-ee60-40a6-af0b-3fa7df1cae1a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1453977923-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c48b059b6f7d49a8ba5ecf28125ddaa9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape873e765-23", "ovs_interfaceid": "e873e765-2324-4273-81ba-e543ebc3b867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 799.416186] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:cd:da', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ab93899c-92b2-4d84-95a6-192234add28c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e873e765-2324-4273-81ba-e543ebc3b867', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.425419] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating folder: Project (c48b059b6f7d49a8ba5ecf28125ddaa9). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.430177] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.431366] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.441223] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f7676fb3-2974-4bfd-827f-72ac5be9a7f0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.442369] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4618ec-51d9-4beb-85af-d26f45eb6eb9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.452762] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2576b9d0-5ce5-460b-8ef7-656933a1f7cc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.459180] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Created folder: Project (c48b059b6f7d49a8ba5ecf28125ddaa9) in parent group-v1000566. [ 799.459416] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating folder: Instances. Parent ref: group-v1000689. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.460473] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c1878800-9a70-4acb-abdf-b8fb8333be36 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.472969] env[66641]: DEBUG nova.compute.provider_tree [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.484278] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Created folder: Instances in parent group-v1000689. [ 799.484541] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 799.484894] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.485184] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a670e7ab-6c0b-49dd-8953-f29624dc4f9b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.506716] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.506716] env[66641]: value = "task-5146051" [ 799.506716] env[66641]: _type = "Task" [ 799.506716] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.515641] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146051, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.556992] env[66641]: DEBUG oslo_concurrency.lockutils [None req-596135bd-a37f-4c72-8284-d0642639040f tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.642s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 799.696857] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b11d4a-aede-4ad3-bfe2-4683ec95bb05 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.706254] env[66641]: DEBUG nova.network.neutron [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 799.732699] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c2ed547-7152-4c7e-b955-80b6e095efec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.754012] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da224948-3a64-4a87-a3b7-fd91d791f06f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.777879] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564c597b-222f-47c0-9303-c358c2ec1de4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.799030] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 67 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 799.883543] env[66641]: WARNING openstack [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 799.883893] env[66641]: WARNING openstack [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 799.889496] env[66641]: DEBUG nova.network.neutron [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 799.969819] env[66641]: DEBUG nova.compute.manager [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Received event network-changed-e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 799.969819] env[66641]: DEBUG nova.compute.manager [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Refreshing instance network info cache due to event network-changed-e873e765-2324-4273-81ba-e543ebc3b867. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 799.969925] env[66641]: DEBUG oslo_concurrency.lockutils [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Acquiring lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.970069] env[66641]: DEBUG oslo_concurrency.lockutils [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Acquired lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 799.970213] env[66641]: DEBUG nova.network.neutron [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Refreshing network info cache for port e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 799.976289] env[66641]: DEBUG nova.scheduler.client.report [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 800.018096] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146051, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.021760] env[66641]: DEBUG nova.network.neutron [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 800.210214] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.251087] env[66641]: DEBUG nova.compute.manager [None req-6963fb5c-e3bf-4de3-8247-8beca3e1d539 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance disappeared during snapshot {{(pid=66641) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4652}} [ 800.307934] env[66641]: WARNING neutronclient.v2_0.client [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.354523] env[66641]: DEBUG nova.network.neutron [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Port cdd4e9c6-89a9-4a63-8850-26be458c02fe binding to destination host cpu-1 is already ACTIVE {{(pid=66641) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3236}} [ 800.473887] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.474537] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.486253] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 800.486933] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 800.492776] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 13.155s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 800.524873] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146051, 'name': CreateVM_Task, 'duration_secs': 0.646609} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.524873] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 800.524873] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.524873] env[66641]: WARNING openstack [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.529931] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.530099] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 800.530437] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 800.530999] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Releasing lock "refresh_cache-207483d3-803c-495b-9b93-6f986f3ca56e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 800.531914] env[66641]: DEBUG nova.compute.manager [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 800.532143] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 800.532925] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb4af6a8-7efa-4b6c-ba55-bca83fdbf0f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.535541] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db9838ab-cb57-4d62-8104-afadd773537d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.544055] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 800.544055] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527979f6-edff-55cf-dd46-38e1b8ee321e" [ 800.544055] env[66641]: _type = "Task" [ 800.544055] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.546178] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 800.553882] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-64b34343-46cb-4bf5-ad8e-777e758bb30d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.571599] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 800.571599] env[66641]: value = "task-5146052" [ 800.571599] env[66641]: _type = "Task" [ 800.571599] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.571922] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527979f6-edff-55cf-dd46-38e1b8ee321e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.583837] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146052, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.653627] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.654532] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.713906] env[66641]: DEBUG nova.compute.manager [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=66641) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:924}} [ 800.714245] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 800.740333] env[66641]: DEBUG nova.compute.manager [None req-6963fb5c-e3bf-4de3-8247-8beca3e1d539 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Found 0 images (rotation: 2) {{(pid=66641) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 800.765925] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 800.766492] env[66641]: WARNING openstack [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 800.889173] env[66641]: DEBUG nova.network.neutron [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updated VIF entry in instance network info cache for port e873e765-2324-4273-81ba-e543ebc3b867. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 800.889615] env[66641]: DEBUG nova.network.neutron [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [{"id": "e873e765-2324-4273-81ba-e543ebc3b867", "address": "fa:16:3e:fb:cd:da", "network": {"id": "3cfa3773-ee60-40a6-af0b-3fa7df1cae1a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1453977923-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c48b059b6f7d49a8ba5ecf28125ddaa9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape873e765-23", "ovs_interfaceid": "e873e765-2324-4273-81ba-e543ebc3b867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 800.996415] env[66641]: DEBUG nova.compute.utils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 800.998574] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 800.998574] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 800.999776] env[66641]: WARNING neutronclient.v2_0.client [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 800.999776] env[66641]: WARNING neutronclient.v2_0.client [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 801.000306] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.001477] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.057407] env[66641]: DEBUG nova.policy [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9d314558a0174bc28ad87c735fec2e61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f7e0123f065d40fe8b51e997ae3f6089', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 801.068341] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527979f6-edff-55cf-dd46-38e1b8ee321e, 'name': SearchDatastore_Task, 'duration_secs': 0.024358} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.071438] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.071794] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.072052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.072230] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 801.072437] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.073079] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a4c92d6-7f97-45ea-9de4-ee8317fb946b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.085617] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146052, 'name': PowerOffVM_Task, 'duration_secs': 0.230528} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.089767] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 801.089957] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 801.090291] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.090457] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.091494] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ce4a1bb8-537d-4bcb-944a-d102dd749081 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.093264] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8321530-0bfb-40ea-95a9-5971f9fcf5e9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.101372] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 801.101372] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cdc900-5386-268f-6938-cd7f31992172" [ 801.101372] env[66641]: _type = "Task" [ 801.101372] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.119731] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cdc900-5386-268f-6938-cd7f31992172, 'name': SearchDatastore_Task, 'duration_secs': 0.016565} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.120694] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6f6671a-2483-4bbf-b68d-5c113b84aaa5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.129197] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 801.129464] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 801.129776] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Deleting the datastore file [datastore1] 207483d3-803c-495b-9b93-6f986f3ca56e {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 801.130205] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90e59d9e-b382-46ed-8868-a98e0bba229e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.135121] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 801.135121] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e919b0-18ea-3803-739d-0712a6cba749" [ 801.135121] env[66641]: _type = "Task" [ 801.135121] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.140454] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for the task: (returnval){ [ 801.140454] env[66641]: value = "task-5146054" [ 801.140454] env[66641]: _type = "Task" [ 801.140454] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.147449] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e919b0-18ea-3803-739d-0712a6cba749, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.161113] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146054, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.331070] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a9c550-5807-4f12-9b38-b57b2e6aedab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.339529] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d81a2b6-6f24-4112-84bf-7877e97018b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.384111] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a694d4-c2c8-4234-9a29-6dbbeac99b99 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.394629] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 801.394629] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 801.394629] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 801.395043] env[66641]: DEBUG oslo_concurrency.lockutils [req-7ac39182-2b9f-43a1-a7c3-afdb0ffeea60 req-5e6a93dc-b814-4ed1-b37f-4c8e6ec78a90 service nova] Releasing lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.400457] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2d8210-ad5b-4d04-8917-d89ee538f29b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.408131] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Successfully created port: 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 801.425578] env[66641]: DEBUG nova.compute.provider_tree [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.515249] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 801.602918] env[66641]: DEBUG nova.compute.manager [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 801.603885] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b109f5e6-6f19-4b75-8ed9-19e3a57f09c5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.648172] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52e919b0-18ea-3803-739d-0712a6cba749, 'name': SearchDatastore_Task, 'duration_secs': 0.018178} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.652630] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 801.653085] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] c95b481a-7956-410d-971c-7d94911230bb/c95b481a-7956-410d-971c-7d94911230bb.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 801.654068] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-665b5484-d1b5-4814-8516-34250fd766b6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.667684] env[66641]: DEBUG oslo_vmware.api [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Task: {'id': task-5146054, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181784} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.669586] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 801.669881] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 801.670130] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.670327] env[66641]: INFO nova.compute.manager [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Took 1.14 seconds to destroy the instance on the hypervisor. [ 801.670587] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 801.671239] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 801.671239] env[66641]: value = "task-5146055" [ 801.671239] env[66641]: _type = "Task" [ 801.671239] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.671479] env[66641]: DEBUG nova.compute.manager [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 801.671537] env[66641]: DEBUG nova.network.neutron [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 801.672128] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 801.672526] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 801.689880] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146055, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.928556] env[66641]: DEBUG nova.scheduler.client.report [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 802.115611] env[66641]: INFO nova.compute.manager [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] instance snapshotting [ 802.123051] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0d49630-c8f7-4a06-bac9-2040bec3a191 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.127986] env[66641]: DEBUG nova.network.neutron [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 802.127986] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.128284] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.161198] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-901aedd7-5b6b-4e42-b3d5-7e9015d5b2f3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.191873] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146055, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.401854] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 802.402320] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 802.529234] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 802.557476] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 802.557781] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 802.557929] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 802.558128] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 802.558290] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 802.558448] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 802.558661] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.558832] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 802.559016] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 802.559180] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 802.559376] env[66641]: DEBUG nova.virt.hardware [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 802.560531] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44a1328a-6681-4c24-9350-c3409913be73 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.569853] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb918d6d-606c-4436-9dd6-f9c3502b2503 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.640360] env[66641]: DEBUG nova.network.neutron [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 802.676246] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 802.676644] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-be7143e2-8351-4a99-b581-41d7c5aa7121 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.687720] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 802.687720] env[66641]: value = "task-5146056" [ 802.687720] env[66641]: _type = "Task" [ 802.687720] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.695713] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146055, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.824517} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.696464] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] c95b481a-7956-410d-971c-7d94911230bb/c95b481a-7956-410d-971c-7d94911230bb.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.696810] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.697154] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f204cb3-1f6b-4b30-9758-bfe8a20d604b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.702562] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146056, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.708326] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 802.708326] env[66641]: value = "task-5146057" [ 802.708326] env[66641]: _type = "Task" [ 802.708326] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.719050] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146057, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.940258] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.448s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 802.945473] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.597s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 802.945783] env[66641]: DEBUG nova.objects.instance [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lazy-loading 'resources' on Instance uuid bc13bbe2-6e02-4d98-9e50-94f772d89ac0 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 802.990519] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Successfully updated port: 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 803.005902] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 803.007847] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23511bc-0fe6-4611-8d67-d64140a8e8a3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.021684] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 803.024118] env[66641]: ERROR oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk due to incomplete transfer. [ 803.025334] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-89c35aef-c35b-4144-88b6-ae894757b418 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.037623] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c8578d-522d-f2cb-4665-92a733593d72/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 803.037780] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Uploaded image 2875e846-84c5-4d52-b64d-3ed1734800e4 to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 803.040239] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 803.040594] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-aa7968f9-873b-4f56-adc4-63fd19073811 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.049721] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.049834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.049970] env[66641]: DEBUG nova.network.neutron [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 803.052869] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 803.052869] env[66641]: value = "task-5146058" [ 803.052869] env[66641]: _type = "Task" [ 803.052869] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.064865] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146058, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.143208] env[66641]: INFO nova.compute.manager [-] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Took 1.47 seconds to deallocate network for instance. [ 803.200418] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146056, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.219493] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146057, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092718} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.219804] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 803.220671] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8462fa4c-cbd3-4e13-b6a6-2460282a446d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.244466] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] c95b481a-7956-410d-971c-7d94911230bb/c95b481a-7956-410d-971c-7d94911230bb.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 803.244817] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-341b7f98-6841-463f-b048-681d5232ec3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.268610] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 803.268610] env[66641]: value = "task-5146059" [ 803.268610] env[66641]: _type = "Task" [ 803.268610] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.278292] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146059, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.370820] env[66641]: DEBUG nova.compute.manager [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 803.371095] env[66641]: DEBUG oslo_concurrency.lockutils [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.371350] env[66641]: DEBUG oslo_concurrency.lockutils [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 803.371552] env[66641]: DEBUG oslo_concurrency.lockutils [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 803.371753] env[66641]: DEBUG nova.compute.manager [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] No waiting events found dispatching network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 803.371955] env[66641]: WARNING nova.compute.manager [req-39d6d223-bf16-4b6f-a155-e87e6ca28e2f req-9b805e31-d639-47e0-a20c-3d95a010766e service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received unexpected event network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 for instance with vm_state building and task_state spawning. [ 803.496559] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.496742] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 803.496931] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 803.557140] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.558156] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 803.575806] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146058, 'name': Destroy_Task, 'duration_secs': 0.38314} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.576745] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Destroyed the VM [ 803.577166] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 803.580821] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9139251d-467b-4293-a46a-335a301e688a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.589829] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 803.589829] env[66641]: value = "task-5146060" [ 803.589829] env[66641]: _type = "Task" [ 803.589829] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.600280] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146060, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.625871] env[66641]: INFO nova.scheduler.client.report [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted allocation for migration 553a2d88-0488-434e-be83-aa85f338f673 [ 803.651397] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 803.711137] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146056, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.783511] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146059, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.811324] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eba3edb-86ee-452b-8203-f476967dab16 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.820378] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94fbacbe-9d40-4079-bc87-eb33a8ff468e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.862163] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-751fd0ef-01e8-4f67-8414-6d62e3b0e073 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.870125] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e38d58c-f649-48db-8607-c9a5ac09cc12 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.886461] env[66641]: DEBUG nova.compute.provider_tree [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.993436] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 803.993819] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.002308] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.002734] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.010651] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 804.106301] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146060, 'name': RemoveSnapshot_Task} progress is 31%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.137172] env[66641]: DEBUG oslo_concurrency.lockutils [None req-128a0b3a-6ea4-40b3-a4e6-7ba91973831c tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 20.775s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.211737] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146056, 'name': CreateSnapshot_Task, 'duration_secs': 1.253609} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.211990] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 804.212839] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d923ae-c5eb-41ca-8518-cafb2afb93d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.285969] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146059, 'name': ReconfigVM_Task, 'duration_secs': 0.66028} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.286313] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Reconfigured VM instance instance-0000002a to attach disk [datastore2] c95b481a-7956-410d-971c-7d94911230bb/c95b481a-7956-410d-971c-7d94911230bb.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 804.288213] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76eba53c-5bad-4d80-8f1e-fe83ac51e9f5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.299395] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 804.299395] env[66641]: value = "task-5146061" [ 804.299395] env[66641]: _type = "Task" [ 804.299395] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.311140] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146061, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.395819] env[66641]: DEBUG nova.scheduler.client.report [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 804.414183] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.414573] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.445320] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.445765] env[66641]: WARNING openstack [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.604609] env[66641]: DEBUG oslo_vmware.api [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146060, 'name': RemoveSnapshot_Task, 'duration_secs': 0.609426} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.608210] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 804.608588] env[66641]: INFO nova.compute.manager [None req-cb38ecca-119a-403b-9972-5608d968ed68 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Took 21.16 seconds to snapshot the instance on the hypervisor. [ 804.687225] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 804.687600] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 804.732966] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 804.733353] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-3311f339-8705-4110-a4e4-2f29d5e37beb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.743164] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 804.743164] env[66641]: value = "task-5146062" [ 804.743164] env[66641]: _type = "Task" [ 804.743164] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.751945] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146062, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.807070] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146061, 'name': Rename_Task, 'duration_secs': 0.392231} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.807360] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 804.807609] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-adecc76a-535d-4dc2-aa5b-705fea32131c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.814911] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 804.814911] env[66641]: value = "task-5146063" [ 804.814911] env[66641]: _type = "Task" [ 804.814911] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.825895] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146063, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.903119] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.958s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 804.907745] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.524s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 804.907745] env[66641]: DEBUG nova.objects.instance [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lazy-loading 'resources' on Instance uuid 05882781-78be-4568-95f4-2fccc4cf4dfe {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.937923] env[66641]: INFO nova.scheduler.client.report [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Deleted allocations for instance bc13bbe2-6e02-4d98-9e50-94f772d89ac0 [ 805.046369] env[66641]: DEBUG nova.network.neutron [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.255624] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146062, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.332600] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146063, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.342373] env[66641]: DEBUG nova.network.neutron [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 805.450460] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d519a6c6-6352-4a6d-81e8-c11fd2e7429d tempest-ServersAdminNegativeTestJSON-864572218 tempest-ServersAdminNegativeTestJSON-864572218-project-member] Lock "bc13bbe2-6e02-4d98-9e50-94f772d89ac0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.366s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 805.549480] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.760125] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146062, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.762354] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141fbe3b-71f4-4951-b082-b0de9320aa41 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.772686] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e5f0e8-33dd-47a5-a482-340cfb32541e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.817310] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf5d1114-1d5a-4735-842b-576ae6e0be34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.821276] env[66641]: DEBUG nova.compute.manager [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-changed-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 805.821528] env[66641]: DEBUG nova.compute.manager [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing instance network info cache due to event network-changed-8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 805.821668] env[66641]: DEBUG oslo_concurrency.lockutils [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.835518] env[66641]: DEBUG oslo_vmware.api [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146063, 'name': PowerOnVM_Task, 'duration_secs': 0.947372} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.835932] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 805.836193] env[66641]: INFO nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 9.89 seconds to spawn the instance on the hypervisor. [ 805.836423] env[66641]: DEBUG nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 805.837659] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10bc31dd-711c-4241-aedd-dad77ada904b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.842574] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997c52ac-4f09-4178-ad99-dc0c2445ad61 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.846518] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 805.846823] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance network_info: |[{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 805.847167] env[66641]: DEBUG oslo_concurrency.lockutils [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 805.847344] env[66641]: DEBUG nova.network.neutron [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 805.848645] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:6f:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bed634b-5119-478e-a71f-a832ef16f035', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 805.856658] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating folder: Project (f7e0123f065d40fe8b51e997ae3f6089). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.868818] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac0950b5-d0ab-422e-b025-a13f11074ec7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.874444] env[66641]: DEBUG nova.compute.provider_tree [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 805.887069] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Created folder: Project (f7e0123f065d40fe8b51e997ae3f6089) in parent group-v1000566. [ 805.887497] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating folder: Instances. Parent ref: group-v1000694. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 805.887965] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d6ce09d-c984-492a-8b9d-399c328eba72 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.899766] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Created folder: Instances in parent group-v1000694. [ 805.900051] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 805.900261] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 805.900481] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-06a1d226-fa54-4f38-8d84-fbd2861e2619 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.923819] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 805.923819] env[66641]: value = "task-5146066" [ 805.923819] env[66641]: _type = "Task" [ 805.923819] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 805.933183] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146066, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.080057] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e9976f-cca0-429e-a350-eb5dd1500c5b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.102456] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35bd3e2b-7fe2-4149-ae40-0eeb726bc830 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.111516] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 83 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 806.263568] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146062, 'name': CloneVM_Task} progress is 95%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.359813] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.360331] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.384490] env[66641]: DEBUG nova.scheduler.client.report [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 806.393145] env[66641]: INFO nova.compute.manager [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 23.24 seconds to build instance. [ 806.438212] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146066, 'name': CreateVM_Task, 'duration_secs': 0.472262} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.438428] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 806.439466] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.439597] env[66641]: WARNING openstack [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.446793] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.446986] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.447316] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 806.448103] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68d33faf-226c-43ef-8c6b-358190e136d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.455562] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 806.455562] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285ef86-01c8-cb4d-a180-0e5707fc7620" [ 806.455562] env[66641]: _type = "Task" [ 806.455562] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.470652] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285ef86-01c8-cb4d-a180-0e5707fc7620, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.618297] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 806.618718] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78fd76a7-22f0-4763-9327-714ce001a36f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.627284] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 806.627284] env[66641]: value = "task-5146067" [ 806.627284] env[66641]: _type = "Task" [ 806.627284] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.638955] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146067, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.685674] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.686073] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.686375] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 806.686718] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.687026] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.690412] env[66641]: INFO nova.compute.manager [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Terminating instance [ 806.730160] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 806.730420] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 806.761113] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146062, 'name': CloneVM_Task, 'duration_secs': 1.854632} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.761421] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Created linked-clone VM from snapshot [ 806.764504] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14b39df5-c4b0-4d95-8922-f11ae15fc3d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.773578] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Uploading image d15990dc-fb1d-4f92-924b-7419d053836a {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 806.807209] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 806.807209] env[66641]: value = "vm-1000693" [ 806.807209] env[66641]: _type = "VirtualMachine" [ 806.807209] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 806.807902] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7e6dd903-5195-4479-b478-f6fd2c8dfc7e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.816332] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lease: (returnval){ [ 806.816332] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231b9fa-74b7-6ff6-7a32-648f4938f70f" [ 806.816332] env[66641]: _type = "HttpNfcLease" [ 806.816332] env[66641]: } obtained for exporting VM: (result){ [ 806.816332] env[66641]: value = "vm-1000693" [ 806.816332] env[66641]: _type = "VirtualMachine" [ 806.816332] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 806.816620] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the lease: (returnval){ [ 806.816620] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231b9fa-74b7-6ff6-7a32-648f4938f70f" [ 806.816620] env[66641]: _type = "HttpNfcLease" [ 806.816620] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 806.824717] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.824717] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231b9fa-74b7-6ff6-7a32-648f4938f70f" [ 806.824717] env[66641]: _type = "HttpNfcLease" [ 806.824717] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 806.893462] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.987s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.896352] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.270s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 806.896547] env[66641]: DEBUG nova.objects.instance [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 806.899715] env[66641]: DEBUG oslo_concurrency.lockutils [None req-b57d7f86-56c6-4c14-97ec-19915202a51e tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.762s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 806.932822] env[66641]: INFO nova.scheduler.client.report [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Deleted allocations for instance 05882781-78be-4568-95f4-2fccc4cf4dfe [ 806.970401] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5285ef86-01c8-cb4d-a180-0e5707fc7620, 'name': SearchDatastore_Task, 'duration_secs': 0.01994} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.970734] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 806.970973] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 806.971377] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 806.971614] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 806.971850] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 806.972100] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-bcca69fb-7efc-47af-8c53-24c53296042b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.984814] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 806.985102] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 806.986249] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722e35bf-ee69-4fcb-8e80-5ec5caaa8729 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.992843] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 806.992843] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9db3a-194c-93b7-402c-4b1ace35634c" [ 806.992843] env[66641]: _type = "Task" [ 806.992843] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.002347] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9db3a-194c-93b7-402c-4b1ace35634c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.072097] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 807.072735] env[66641]: WARNING openstack [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 807.140275] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146067, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.199217] env[66641]: DEBUG nova.compute.manager [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 807.199217] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 807.199217] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8565c7c8-4bb7-46c0-aff3-c6d87b6697b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.207956] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 807.208354] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aabde422-b97f-40df-9ebe-4b1fa51f0fb8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.215081] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 807.215081] env[66641]: value = "task-5146069" [ 807.215081] env[66641]: _type = "Task" [ 807.215081] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.232671] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5146069, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.330024] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 807.330024] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231b9fa-74b7-6ff6-7a32-648f4938f70f" [ 807.330024] env[66641]: _type = "HttpNfcLease" [ 807.330024] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 807.330024] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 807.330024] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231b9fa-74b7-6ff6-7a32-648f4938f70f" [ 807.330024] env[66641]: _type = "HttpNfcLease" [ 807.330024] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 807.333225] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55c8494a-d986-4328-868f-0f507336e113 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.340020] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 807.340333] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 807.412521] env[66641]: DEBUG nova.network.neutron [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updated VIF entry in instance network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 807.412679] env[66641]: DEBUG nova.network.neutron [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 807.440696] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f4d77fae-0244-4aa1-9f74-e946d324c92c tempest-ImagesTestJSON-964234499 tempest-ImagesTestJSON-964234499-project-member] Lock "05882781-78be-4568-95f4-2fccc4cf4dfe" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.684s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.464110] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2c6db99a-aebd-42ac-b371-bce359f029a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.507372] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a9db3a-194c-93b7-402c-4b1ace35634c, 'name': SearchDatastore_Task, 'duration_secs': 0.010967} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.508207] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd389a0f-3207-42fa-a9ac-75ee9979777d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.514761] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 807.514761] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ab05ce-5779-4882-79a4-b93e69cd4c69" [ 807.514761] env[66641]: _type = "Task" [ 807.514761] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.525943] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ab05ce-5779-4882-79a4-b93e69cd4c69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.643785] env[66641]: DEBUG oslo_vmware.api [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146067, 'name': PowerOnVM_Task, 'duration_secs': 0.613357} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.644056] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 807.645716] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c9de7c0b-fc6c-4bd9-825d-4f080ae50f80 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance '48f99287-b737-45fa-ad59-9e1425afa3d5' progress to 100 {{(pid=66641) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 807.680402] env[66641]: DEBUG nova.compute.manager [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 807.681116] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6896e3e-1739-43d1-a2b6-6925ea42ab08 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.730628] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5146069, 'name': PowerOffVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.917397] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b540b38-a2fc-4eea-8890-ba6265fa7f23 tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 807.917397] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.581s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 807.917397] env[66641]: DEBUG nova.objects.instance [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lazy-loading 'resources' on Instance uuid dad5ed23-71a1-4b55-856f-2484f8e62708 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 807.918201] env[66641]: DEBUG oslo_concurrency.lockutils [req-0d323c1e-bd61-4feb-9f73-2e250ef08370 req-b01352e3-5b02-4a34-bd1e-fc80dfacd954 service nova] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.038980] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ab05ce-5779-4882-79a4-b93e69cd4c69, 'name': SearchDatastore_Task, 'duration_secs': 0.010628} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.039396] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 808.039735] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 808.040077] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8877c6fb-10a6-485f-a202-0816984fd2b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.051782] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 808.051782] env[66641]: value = "task-5146070" [ 808.051782] env[66641]: _type = "Task" [ 808.051782] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.062246] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146070, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.194861] env[66641]: INFO nova.compute.manager [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] instance snapshotting [ 808.200021] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c3eeb02-680a-49c6-82a0-26c5fb20407c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.238158] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b6c332-261b-496f-b391-40ce7965a758 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.255021] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5146069, 'name': PowerOffVM_Task, 'duration_secs': 0.516173} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.257588] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 808.257908] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 808.260874] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8f81b87b-3386-4036-83d8-7dc866199de8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.360717] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 808.360717] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 808.360717] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleting the datastore file [datastore1] 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 808.361878] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40661f83-dfcf-43a9-84be-630f1aedba60 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.371877] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for the task: (returnval){ [ 808.371877] env[66641]: value = "task-5146072" [ 808.371877] env[66641]: _type = "Task" [ 808.371877] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.385166] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5146072, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.564452] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146070, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.767623] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 808.768595] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f96cd1b0-6ac8-4fde-989f-ea6a1d1e8e80 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.781132] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 808.781132] env[66641]: value = "task-5146073" [ 808.781132] env[66641]: _type = "Task" [ 808.781132] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.787730] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07165d44-343f-4e82-bafd-6791b9668e12 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.799641] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146073, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.806542] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a55f61f-7ade-49e4-a8e3-f8cc4c92a992 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.846322] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-babba23e-cee5-4601-873b-2c3333535f26 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.855589] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9763388-71d5-4dd7-aa08-6ab0cd25cf26 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.871665] env[66641]: DEBUG nova.compute.provider_tree [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.884407] env[66641]: DEBUG oslo_vmware.api [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Task: {'id': task-5146072, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.399986} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.884732] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 808.884946] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 808.885275] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.885510] env[66641]: INFO nova.compute.manager [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Took 1.69 seconds to destroy the instance on the hypervisor. [ 808.886836] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 808.886836] env[66641]: DEBUG nova.compute.manager [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 808.886836] env[66641]: DEBUG nova.network.neutron [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 808.886836] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.887358] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 808.997162] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 808.997162] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.064399] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146070, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659229} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.065977] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 809.066270] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.067578] env[66641]: DEBUG nova.compute.manager [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Received event network-changed-e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 809.068013] env[66641]: DEBUG nova.compute.manager [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Refreshing instance network info cache due to event network-changed-e873e765-2324-4273-81ba-e543ebc3b867. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 809.068314] env[66641]: DEBUG oslo_concurrency.lockutils [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Acquiring lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.068512] env[66641]: DEBUG oslo_concurrency.lockutils [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Acquired lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 809.068720] env[66641]: DEBUG nova.network.neutron [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Refreshing network info cache for port e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 809.070126] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d368d64e-7e2f-4f00-a5ba-bec8f532f353 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.080034] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 809.080034] env[66641]: value = "task-5146074" [ 809.080034] env[66641]: _type = "Task" [ 809.080034] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.093276] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146074, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.293190] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146073, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.375392] env[66641]: DEBUG nova.scheduler.client.report [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 809.575008] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.575516] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.597145] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146074, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08089} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.598363] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.599543] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f27a59c7-5ad6-43c1-b9f3-9a2e2fbfbdaa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.632327] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.633633] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5bf1170b-bbb9-44df-9a22-3f17310b5250 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.663793] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 809.663793] env[66641]: value = "task-5146075" [ 809.663793] env[66641]: _type = "Task" [ 809.663793] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.675718] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146075, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.765324] env[66641]: DEBUG nova.network.neutron [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 809.795614] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146073, 'name': CreateSnapshot_Task, 'duration_secs': 0.897183} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.795959] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 809.796774] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de87c78-5e25-4d37-b89c-5dbc554cf37b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.816990] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.817482] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.882239] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.965s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 809.885637] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.007s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 809.887330] env[66641]: INFO nova.compute.claims [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.913532] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 809.914302] env[66641]: WARNING openstack [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 809.925255] env[66641]: INFO nova.scheduler.client.report [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Deleted allocations for instance dad5ed23-71a1-4b55-856f-2484f8e62708 [ 810.155774] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "acde41f9-4256-4e63-98e3-fd092c66c71e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.155774] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.177648] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146075, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.214254] env[66641]: DEBUG nova.network.neutron [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updated VIF entry in instance network info cache for port e873e765-2324-4273-81ba-e543ebc3b867. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 810.214386] env[66641]: DEBUG nova.network.neutron [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [{"id": "e873e765-2324-4273-81ba-e543ebc3b867", "address": "fa:16:3e:fb:cd:da", "network": {"id": "3cfa3773-ee60-40a6-af0b-3fa7df1cae1a", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1453977923-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.168", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c48b059b6f7d49a8ba5ecf28125ddaa9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ab93899c-92b2-4d84-95a6-192234add28c", "external-id": "nsx-vlan-transportzone-697", "segmentation_id": 697, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape873e765-23", "ovs_interfaceid": "e873e765-2324-4273-81ba-e543ebc3b867", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 810.269617] env[66641]: INFO nova.compute.manager [-] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Took 1.38 seconds to deallocate network for instance. [ 810.324463] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 810.325830] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-e9199fdf-5a46-41ed-99fa-37953e466e1d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.339611] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 810.339611] env[66641]: value = "task-5146076" [ 810.339611] env[66641]: _type = "Task" [ 810.339611] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.350467] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146076, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.444926] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d6ad2d97-54f6-4b99-95ba-b69cecb4d0ab tempest-ServerRescueTestJSONUnderV235-1185073902 tempest-ServerRescueTestJSONUnderV235-1185073902-project-member] Lock "dad5ed23-71a1-4b55-856f-2484f8e62708" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.702s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 810.583380] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.583694] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 810.583894] env[66641]: DEBUG nova.compute.manager [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Going to confirm migration 4 {{(pid=66641) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:5307}} [ 810.662786] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 810.685014] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146075, 'name': ReconfigVM_Task, 'duration_secs': 0.711614} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.685014] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Reconfigured VM instance instance-0000002b to attach disk [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.685014] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a5fd0c3-2591-4557-a9a8-1546d410c857 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.697363] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 810.697363] env[66641]: value = "task-5146077" [ 810.697363] env[66641]: _type = "Task" [ 810.697363] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.710427] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146077, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.718498] env[66641]: DEBUG oslo_concurrency.lockutils [req-adae6c66-b668-41d0-84db-1eec2c6e364b req-c4961637-0f4d-4648-bcaa-cc13cd5b873c service nova] Releasing lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 810.776769] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 810.851848] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146076, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.092633] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.093080] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.161668] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 811.162497] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 811.169128] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.169313] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquired lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 811.169493] env[66641]: DEBUG nova.network.neutron [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 811.169674] env[66641]: DEBUG nova.objects.instance [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lazy-loading 'info_cache' on Instance uuid 48f99287-b737-45fa-ad59-9e1425afa3d5 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 811.198326] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231d231f-21d8-4072-88dd-7c8e24448b18 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.202241] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 811.212671] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146077, 'name': Rename_Task, 'duration_secs': 0.356845} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.215070] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 811.215413] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8b1821b4-264b-43ac-9afe-6f4fb8d028c5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.217993] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2363862-6f4b-401e-8418-fc1686ff5644 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.268360] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f05571ba-73a0-400f-a577-a17d5033f5df {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.273295] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 811.273295] env[66641]: value = "task-5146078" [ 811.273295] env[66641]: _type = "Task" [ 811.273295] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.281352] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16973a6e-d812-4260-ac0e-9ea3af7edb35 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.289634] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146078, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.301975] env[66641]: DEBUG nova.compute.provider_tree [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.352652] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146076, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.787829] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146078, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.805296] env[66641]: DEBUG nova.scheduler.client.report [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 811.854172] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146076, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.005545] env[66641]: DEBUG nova.compute.manager [req-59a67cbf-5692-47df-8cd1-e7422dbf4069 req-8f27c429-0305-4e5d-a335-bc65b6759f3b service nova] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Received event network-vif-deleted-2f5f78fa-cbb1-4890-8d2a-68ce8f5bca4a {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 812.185085] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.185984] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.284994] env[66641]: DEBUG oslo_vmware.api [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146078, 'name': PowerOnVM_Task, 'duration_secs': 0.787794} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.285317] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.285539] env[66641]: INFO nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Took 9.76 seconds to spawn the instance on the hypervisor. [ 812.285746] env[66641]: DEBUG nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 812.286582] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ca79f2b-c43e-42d1-93dd-1a3c98ccbb90 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.315973] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.430s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 812.315973] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 812.321675] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.007s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 812.321942] env[66641]: DEBUG nova.objects.instance [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lazy-loading 'resources' on Instance uuid 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 812.355938] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146076, 'name': CloneVM_Task, 'duration_secs': 1.85445} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.356333] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Created linked-clone VM from snapshot [ 812.357150] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d7b1f46-e048-4101-9dae-ee2700812240 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.368751] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Uploading image 502c337b-699c-4132-a047-33d44a1bd1bd {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 812.405265] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 812.405265] env[66641]: value = "vm-1000698" [ 812.405265] env[66641]: _type = "VirtualMachine" [ 812.405265] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 812.405564] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-a9ca5a6f-d651-4a5c-9f17-12ecdb2cf2f9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.417825] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease: (returnval){ [ 812.417825] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52814a53-a22b-9816-5930-ab0762f3e34e" [ 812.417825] env[66641]: _type = "HttpNfcLease" [ 812.417825] env[66641]: } obtained for exporting VM: (result){ [ 812.417825] env[66641]: value = "vm-1000698" [ 812.417825] env[66641]: _type = "VirtualMachine" [ 812.417825] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 812.418232] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the lease: (returnval){ [ 812.418232] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52814a53-a22b-9816-5930-ab0762f3e34e" [ 812.418232] env[66641]: _type = "HttpNfcLease" [ 812.418232] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 812.426513] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.426513] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52814a53-a22b-9816-5930-ab0762f3e34e" [ 812.426513] env[66641]: _type = "HttpNfcLease" [ 812.426513] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 812.807549] env[66641]: INFO nova.compute.manager [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Took 26.89 seconds to build instance. [ 812.828989] env[66641]: DEBUG nova.compute.utils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 812.831788] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 812.831993] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 812.832330] env[66641]: WARNING neutronclient.v2_0.client [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.832624] env[66641]: WARNING neutronclient.v2_0.client [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 812.833243] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 812.833593] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 812.928659] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 812.928659] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52814a53-a22b-9816-5930-ab0762f3e34e" [ 812.928659] env[66641]: _type = "HttpNfcLease" [ 812.928659] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 812.931447] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 812.931447] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52814a53-a22b-9816-5930-ab0762f3e34e" [ 812.931447] env[66641]: _type = "HttpNfcLease" [ 812.931447] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 812.932731] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d0243b0-9d0f-4951-ab44-b0ebe61e411f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.941541] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 812.941743] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 813.093022] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-24f24f53-e5f8-4729-8a4d-3fabb381be5e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.145641] env[66641]: DEBUG nova.policy [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e3991023cd3c4b75a5f1f09bb8ca8b8f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3a9466d8e6fa49c59a14110b417664c6', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 813.149688] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c716d71-434a-4374-9e1e-461c61e2234e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.159816] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f70424-d691-4d50-8ed6-33b03928a50a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.196508] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 813.196885] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 813.204415] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce7767c9-1864-4c63-b155-289bedf83ad1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.213570] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6df7bf2-8668-42e1-b542-d4a2fc003600 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.233938] env[66641]: DEBUG nova.compute.provider_tree [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.310223] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83c8e2f5-d1d6-4580-a6e4-174a0a7b4c04 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.418s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 813.332299] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 813.738735] env[66641]: DEBUG nova.scheduler.client.report [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 814.153774] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Successfully created port: 049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 814.220490] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 814.221022] env[66641]: WARNING openstack [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 814.245141] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.923s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 814.247968] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.274s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 814.248487] env[66641]: DEBUG nova.objects.instance [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lazy-loading 'resources' on Instance uuid 081457ae-e152-410c-bca7-4d43b95eee10 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.289185] env[66641]: INFO nova.scheduler.client.report [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Deleted allocations for instance 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db [ 814.348532] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 814.806032] env[66641]: DEBUG oslo_concurrency.lockutils [None req-12e84625-f27d-4281-9c3b-c1a7bcdf4424 tempest-ServerGroupTestJSON-2090438877 tempest-ServerGroupTestJSON-2090438877-project-member] Lock "34a98372-2ab7-4b21-8a0e-2fc3b91ef4db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.682s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 815.093467] env[66641]: DEBUG nova.network.neutron [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [{"id": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "address": "fa:16:3e:9d:c2:42", "network": {"id": "8a78905f-dde5-493a-902c-2092e9d9853b", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1670577956-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "b2c8481015524aee95a933f61082faec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4e2ec358-9bc5-4dd6-8f4e-0d6ec225282a", "external-id": "nsx-vlan-transportzone-843", "segmentation_id": 843, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdd4e9c6-89", "ovs_interfaceid": "cdd4e9c6-89a9-4a63-8850-26be458c02fe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 815.168031] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0285398-985a-43e5-94bd-20c0302425fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.184071] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-309458be-9292-4d3d-94a3-1dad2fdfd4bf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.227024] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596ef073-a99e-4029-a17e-1054f8f2eaa9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.236844] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac415a9-f1a6-47c7-92d8-15ce9706651c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.253185] env[66641]: DEBUG nova.compute.provider_tree [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.599685] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Releasing lock "refresh_cache-48f99287-b737-45fa-ad59-9e1425afa3d5" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 815.599966] env[66641]: DEBUG nova.objects.instance [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lazy-loading 'migration_context' on Instance uuid 48f99287-b737-45fa-ad59-9e1425afa3d5 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.759795] env[66641]: DEBUG nova.scheduler.client.report [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 815.948955] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Successfully updated port: 049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 816.103765] env[66641]: DEBUG nova.objects.base [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Object Instance<48f99287-b737-45fa-ad59-9e1425afa3d5> lazy-loaded attributes: info_cache,migration_context {{(pid=66641) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 816.104885] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-792f72bb-b652-44df-a6f3-f4ed17641d57 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.130614] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a1becfdd-9f64-4030-9c27-6206ca46bc97 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.138546] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 816.138546] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5226decc-b7c2-46d4-92d3-a00737278424" [ 816.138546] env[66641]: _type = "Task" [ 816.138546] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.149135] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5226decc-b7c2-46d4-92d3-a00737278424, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.264896] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.017s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.267881] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 15.553s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 816.317378] env[66641]: INFO nova.scheduler.client.report [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Deleted allocations for instance 081457ae-e152-410c-bca7-4d43b95eee10 [ 816.378704] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 816.378893] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 816.379070] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 816.379260] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 816.379402] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 816.379543] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 816.379756] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.379905] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 816.380302] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 816.380647] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 816.380867] env[66641]: DEBUG nova.virt.hardware [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 816.383698] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f565382f-c82c-4dbf-9d6a-db9d66a8ed90 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.398514] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 816.400283] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04e9d3b-f783-4ef3-b396-14f2c31a8810 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.406272] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b75e242a-5587-47c0-b4a1-44c418dc3602 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.425347] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 816.425533] env[66641]: ERROR oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk due to incomplete transfer. [ 816.426466] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-a029f9ea-dc5b-4e67-b8a1-e1bf03d68e9b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.436793] env[66641]: DEBUG oslo_vmware.rw_handles [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5213c3b2-7b5f-2379-3cb1-bee14f368e53/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 816.437019] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Uploaded image d15990dc-fb1d-4f92-924b-7419d053836a to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 816.439410] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 816.439754] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-af9c07d5-07ad-4d53-8337-816479c7fd9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.449208] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 816.449208] env[66641]: value = "task-5146080" [ 816.449208] env[66641]: _type = "Task" [ 816.449208] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.461326] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.461326] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquired lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 816.461326] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 816.463919] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146080, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.662572] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5226decc-b7c2-46d4-92d3-a00737278424, 'name': SearchDatastore_Task, 'duration_secs': 0.011609} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.664488] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 816.694454] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.695462] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 816.776356] env[66641]: DEBUG nova.objects.instance [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lazy-loading 'migration_context' on Instance uuid 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.827075] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1decd29b-d347-4b97-9056-a92f4456d2dd tempest-ServersListShow2100Test-293155119 tempest-ServersListShow2100Test-293155119-project-member] Lock "081457ae-e152-410c-bca7-4d43b95eee10" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.903s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 816.966695] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 816.966695] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 816.971248] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 816.973081] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146080, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.006105] env[66641]: DEBUG nova.compute.manager [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received event network-vif-plugged-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 817.006305] env[66641]: DEBUG oslo_concurrency.lockutils [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.006497] env[66641]: DEBUG oslo_concurrency.lockutils [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] Lock "88ae00a2-6139-4258-b316-0f75032275ec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.006577] env[66641]: DEBUG oslo_concurrency.lockutils [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] Lock "88ae00a2-6139-4258-b316-0f75032275ec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 817.006793] env[66641]: DEBUG nova.compute.manager [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] No waiting events found dispatching network-vif-plugged-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 817.007103] env[66641]: WARNING nova.compute.manager [req-4a29317f-634d-496d-ae5b-4b6f1d39d214 req-b700a495-fe19-41ef-8357-4f4589b88779 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received unexpected event network-vif-plugged-049d8c0f-cda0-4cd9-b972-188efe21a140 for instance with vm_state building and task_state spawning. [ 817.150585] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.151074] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.208381] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.209029] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 817.209029] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 817.468900] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146080, 'name': Destroy_Task, 'duration_secs': 0.762992} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.469235] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Destroyed the VM [ 817.469551] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 817.469752] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-2d853512-c326-44ef-ac2e-f0c75834ce3f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.480736] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 817.480736] env[66641]: value = "task-5146081" [ 817.480736] env[66641]: _type = "Task" [ 817.480736] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.494707] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146081, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.554777] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e91f9463-d1c7-4dc2-8cad-6a26b96336f9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.572426] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9dbab0-4148-4f17-a190-6cb3eb14bb9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.635230] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6830f7a4-0ae4-4523-92cb-01ca401d1d2b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.642174] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.642174] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 817.649438] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7309d9bd-cc01-4100-88f7-cd90f4ebcffa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.669863] env[66641]: DEBUG nova.compute.provider_tree [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.713803] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 817.757500] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 817.757906] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 817.915567] env[66641]: DEBUG nova.network.neutron [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [{"id": "049d8c0f-cda0-4cd9-b972-188efe21a140", "address": "fa:16:3e:15:f5:a4", "network": {"id": "28059287-894a-40a1-8385-d1fddec6f422", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-167219310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a9466d8e6fa49c59a14110b417664c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap049d8c0f-cd", "ovs_interfaceid": "049d8c0f-cda0-4cd9-b972-188efe21a140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 817.992338] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146081, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.144486] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 818.174558] env[66641]: DEBUG nova.scheduler.client.report [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 818.418120] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Releasing lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 818.418478] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Instance network_info: |[{"id": "049d8c0f-cda0-4cd9-b972-188efe21a140", "address": "fa:16:3e:15:f5:a4", "network": {"id": "28059287-894a-40a1-8385-d1fddec6f422", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-167219310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a9466d8e6fa49c59a14110b417664c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap049d8c0f-cd", "ovs_interfaceid": "049d8c0f-cda0-4cd9-b972-188efe21a140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 818.419074] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:15:f5:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '225b6979-9329-403b-91fa-138bd41f6e83', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '049d8c0f-cda0-4cd9-b972-188efe21a140', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 818.432690] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Creating folder: Project (3a9466d8e6fa49c59a14110b417664c6). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.433073] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a1a21b85-e10c-46df-bbc3-2e31a592bcff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.446628] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Created folder: Project (3a9466d8e6fa49c59a14110b417664c6) in parent group-v1000566. [ 818.449907] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Creating folder: Instances. Parent ref: group-v1000699. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 818.449907] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4190adbe-01c7-4571-bde2-b88d7d3a9994 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.458556] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Created folder: Instances in parent group-v1000699. [ 818.458886] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 818.459022] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 818.459278] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d2ebf763-d779-4b81-b777-5290b87f993c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.482163] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 818.482163] env[66641]: value = "task-5146084" [ 818.482163] env[66641]: _type = "Task" [ 818.482163] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.495267] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146081, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.499195] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146084, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.670770] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 818.996226] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146084, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.000062] env[66641]: DEBUG oslo_vmware.api [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146081, 'name': RemoveSnapshot_Task, 'duration_secs': 1.455072} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.000353] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 819.000587] env[66641]: INFO nova.compute.manager [None req-0bfbaee7-069d-413b-84a5-7a61ce4bfdb8 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 16.88 seconds to snapshot the instance on the hypervisor. [ 819.188353] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.920s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 819.196470] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.545s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 819.196738] env[66641]: DEBUG nova.objects.instance [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lazy-loading 'resources' on Instance uuid 207483d3-803c-495b-9b93-6f986f3ca56e {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.442575] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "4e70780a-62ca-4e4a-9366-00dc35750c61" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 819.442963] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 819.501669] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146084, 'name': CreateVM_Task, 'duration_secs': 0.522369} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.502065] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 819.503280] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 819.503970] env[66641]: WARNING openstack [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 819.513100] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.513813] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.514367] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 819.518162] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-244b546b-e1c4-4f2c-97c5-7c7cf21a72d3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.528799] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 819.528799] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52df4264-5af1-1834-6be7-cade26b33b67" [ 819.528799] env[66641]: _type = "Task" [ 819.528799] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.548925] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52df4264-5af1-1834-6be7-cade26b33b67, 'name': SearchDatastore_Task, 'duration_secs': 0.014304} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.549946] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 819.550304] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 819.550433] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.550574] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 819.550749] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 819.551319] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-891f0e78-a71f-44fd-972d-3e837f837ca8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.562256] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 819.562440] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 819.563225] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5b8d2a7-02e7-4b04-b53c-cad60bb7737e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.570633] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 819.570633] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bdbd48-a91c-81f1-3378-d9da6608f55e" [ 819.570633] env[66641]: _type = "Task" [ 819.570633] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.581858] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bdbd48-a91c-81f1-3378-d9da6608f55e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.924505] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5668fc5-8589-4167-9578-ecfd3e106b27 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.931230] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c29f5843-3069-4d94-8591-2d3d601751d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.965652] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 819.972214] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27f0e07e-f739-4f4e-996d-7c193fac27e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.979115] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2abc1676-e98b-4798-8304-9f1bd3128afe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.994869] env[66641]: DEBUG nova.compute.provider_tree [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.083684] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bdbd48-a91c-81f1-3378-d9da6608f55e, 'name': SearchDatastore_Task, 'duration_secs': 0.011545} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.084696] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5075c3a7-d9db-4b1e-8eb3-9550cb34c782 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.091804] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 820.091804] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520fe1c5-2669-ffde-4b5c-31739c4cfd3e" [ 820.091804] env[66641]: _type = "Task" [ 820.091804] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.102281] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520fe1c5-2669-ffde-4b5c-31739c4cfd3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.297698] env[66641]: DEBUG nova.compute.manager [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 820.297935] env[66641]: DEBUG nova.compute.manager [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing instance network info cache due to event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 820.298199] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Acquiring lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.298369] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Acquired lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 820.298575] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 820.473236] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "29f03150-f0a2-4424-b758-32891730923c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.473485] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 820.491601] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 820.498713] env[66641]: DEBUG nova.scheduler.client.report [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 820.606281] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]520fe1c5-2669-ffde-4b5c-31739c4cfd3e, 'name': SearchDatastore_Task, 'duration_secs': 0.01297} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.606654] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 820.606901] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 88ae00a2-6139-4258-b316-0f75032275ec/88ae00a2-6139-4258-b316-0f75032275ec.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 820.607228] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e44bcc3f-4482-4b4f-972b-a640e991e707 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.615404] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 820.615404] env[66641]: value = "task-5146085" [ 820.615404] env[66641]: _type = "Task" [ 820.615404] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.624141] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146085, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.736691] env[66641]: INFO nova.compute.manager [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Swapping old allocation on dict_keys(['750ffd2d-5e46-4240-a614-995f2be7c9cb']) held by migration b4026a95-a6cb-4f08-a135-728567b9bd8c for instance [ 820.764579] env[66641]: DEBUG nova.scheduler.client.report [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Overwriting current allocation {'allocations': {'750ffd2d-5e46-4240-a614-995f2be7c9cb': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 65}}, 'project_id': '6624f269bb8840e98c10259f006ce962', 'user_id': 'dfb62f2b801a4c7a899ac65c05c382f1', 'consumer_generation': 1} on consumer 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 {{(pid=66641) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2036}} [ 820.802389] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.803117] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.833155] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.833155] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.878386] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.878553] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquired lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 820.878729] env[66641]: DEBUG nova.network.neutron [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 820.922921] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.923673] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 820.976847] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 820.996142] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 820.996713] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.006345] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.008726] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.232s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.010922] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.010922] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.809s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 821.012951] env[66641]: INFO nova.compute.claims [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.032342] env[66641]: INFO nova.scheduler.client.report [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Deleted allocations for instance 207483d3-803c-495b-9b93-6f986f3ca56e [ 821.041602] env[66641]: INFO nova.scheduler.client.report [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Deleted allocations for instance 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f [ 821.117172] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updated VIF entry in instance network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 821.117677] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [{"id": "049d8c0f-cda0-4cd9-b972-188efe21a140", "address": "fa:16:3e:15:f5:a4", "network": {"id": "28059287-894a-40a1-8385-d1fddec6f422", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-167219310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a9466d8e6fa49c59a14110b417664c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap049d8c0f-cd", "ovs_interfaceid": "049d8c0f-cda0-4cd9-b972-188efe21a140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 821.135228] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146085, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.382190] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.382541] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.488996] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.489394] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.508369] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 821.547814] env[66641]: DEBUG oslo_concurrency.lockutils [None req-f39974f1-9eab-42e1-838b-b68c835d50e7 tempest-ServersAaction247Test-1786006168 tempest-ServersAaction247Test-1786006168-project-member] Lock "207483d3-803c-495b-9b93-6f986f3ca56e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.681s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.550614] env[66641]: DEBUG oslo_concurrency.lockutils [None req-28edf126-06c7-47d6-b4c4-6c06adcfd2cb tempest-ServerDiskConfigTestJSON-1859494360 tempest-ServerDiskConfigTestJSON-1859494360-project-member] Lock "047a5c42-3930-4e6a-b3a5-5dbf55d44a4f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.865s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 821.569583] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 821.569951] env[66641]: WARNING openstack [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 821.623271] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Releasing lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 821.623678] env[66641]: DEBUG nova.compute.manager [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-changed-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 821.623936] env[66641]: DEBUG nova.compute.manager [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing instance network info cache due to event network-changed-8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 821.624114] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.624266] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 821.624427] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 821.632680] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146085, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.531835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.632878] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] 88ae00a2-6139-4258-b316-0f75032275ec/88ae00a2-6139-4258-b316-0f75032275ec.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 821.633139] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 821.634035] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-812ce2b4-6503-4ee7-b01f-110dda86f011 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.641428] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 821.641428] env[66641]: value = "task-5146086" [ 821.641428] env[66641]: _type = "Task" [ 821.641428] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.653022] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146086, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.655853] env[66641]: DEBUG nova.network.neutron [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [{"id": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "address": "fa:16:3e:e8:05:57", "network": {"id": "b08de140-1bf6-41d1-b4d1-3c1eb85d4a1e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.180", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "dde1b7d490614e5b8332835e29fc0c01", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "193994c7-8e1b-4f25-a4a4-d0563845eb28", "external-id": "nsx-vlan-transportzone-607", "segmentation_id": 607, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa44fab6d-1a", "ovs_interfaceid": "a44fab6d-1a8a-4a32-93c1-ebfd24e3d021", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 822.128498] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.128970] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.153655] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146086, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076623} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.156708] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 822.158490] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f499f7f-f16c-45b1-a896-246c2139ecdf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.161438] env[66641]: DEBUG oslo_concurrency.lockutils [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Releasing lock "refresh_cache-58fefaa4-0b17-408f-9329-78f8b5cf3fa7" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 822.161924] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 822.162182] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bfd64c78-9203-43e1-a7d8-2a10e9d59274 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.189110] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 88ae00a2-6139-4258-b316-0f75032275ec/88ae00a2-6139-4258-b316-0f75032275ec.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 822.200510] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed6acaa9-f2e9-44e5-8a2e-c5e6384f9dc8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.217870] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 822.217870] env[66641]: value = "task-5146087" [ 822.217870] env[66641]: _type = "Task" [ 822.217870] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.225949] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 822.225949] env[66641]: value = "task-5146088" [ 822.225949] env[66641]: _type = "Task" [ 822.225949] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.229717] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146087, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.241020] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146088, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.323426] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8810018-2843-4766-b0ef-0e344f95591c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.334484] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82de4445-e85a-4841-a82d-5555c68eba98 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.383277] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 822.383987] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 822.394522] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a8f665f-8548-49ed-9eff-676ca7d6640b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.406635] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae36ad5-668f-4572-9cd7-2afec2fbb0c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.426588] env[66641]: DEBUG nova.compute.provider_tree [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.544288] env[66641]: DEBUG nova.compute.manager [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 822.545313] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da804275-ce39-4859-9d59-0c1b75de2950 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.734384] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146087, 'name': PowerOffVM_Task, 'duration_secs': 0.31414} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.738264] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 822.738986] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:51:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ed028205-0b7e-4e53-9191-4b8a2662b934',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-192737414',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 822.739306] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 822.739468] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 822.739656] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 822.739799] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 822.739942] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 822.740161] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 822.740311] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 822.740472] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 822.740652] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 822.740824] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 822.746696] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01ffa454-b16f-4f29-956d-489e640d5203 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.763407] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146088, 'name': ReconfigVM_Task, 'duration_secs': 0.321305} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.764997] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 88ae00a2-6139-4258-b316-0f75032275ec/88ae00a2-6139-4258-b316-0f75032275ec.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 822.765777] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 822.765777] env[66641]: value = "task-5146089" [ 822.765777] env[66641]: _type = "Task" [ 822.765777] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.766141] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7e358d0e-ad52-4c10-8ad6-2a39972082a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.778193] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146089, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.779935] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 822.779935] env[66641]: value = "task-5146090" [ 822.779935] env[66641]: _type = "Task" [ 822.779935] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.789993] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146090, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.930396] env[66641]: DEBUG nova.scheduler.client.report [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 823.058362] env[66641]: INFO nova.compute.manager [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] instance snapshotting [ 823.061947] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6568443-f613-47e8-a67f-79ff17a4993d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.087567] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb47237c-aefe-4f2d-9ca0-a162a7e78786 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.279778] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146089, 'name': ReconfigVM_Task, 'duration_secs': 0.172212} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.280999] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb42030-f4ec-4c7f-a4a3-80c7ef24e1fa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.295533] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146090, 'name': Rename_Task, 'duration_secs': 0.272577} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.314219] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 823.315256] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:51:09Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='ed028205-0b7e-4e53-9191-4b8a2662b934',id=35,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-192737414',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 823.315664] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 823.316060] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 823.316622] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 823.316844] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 823.317199] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 823.317949] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 823.318260] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 823.318544] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 823.318908] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 823.319266] env[66641]: DEBUG nova.virt.hardware [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 823.320175] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa9f132b-459c-48eb-a21f-8a9d71f335c6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.322107] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4596e46b-3772-4f72-9eb9-b13b65a5bb67 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.329259] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 823.329259] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233d0c1-c73d-db81-e478-ca534eb3ed2a" [ 823.329259] env[66641]: _type = "Task" [ 823.329259] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.334304] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 823.334304] env[66641]: value = "task-5146091" [ 823.334304] env[66641]: _type = "Task" [ 823.334304] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.342215] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233d0c1-c73d-db81-e478-ca534eb3ed2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.349136] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146091, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.437015] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 823.438111] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 823.442765] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.780s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 823.603871] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 823.603871] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-c7f94962-eb6c-47bc-a4a7-6711ef8f4489 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.613452] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 823.613452] env[66641]: value = "task-5146092" [ 823.613452] env[66641]: _type = "Task" [ 823.613452] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.627941] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146092, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.693882] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.694021] env[66641]: WARNING openstack [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.847062] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5233d0c1-c73d-db81-e478-ca534eb3ed2a, 'name': SearchDatastore_Task, 'duration_secs': 0.011022} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.856564] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfiguring VM instance instance-0000001e to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 823.861705] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85489ea0-6e50-4b43-bcb6-57e0d136a7f6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.876725] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146091, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.883187] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 823.883187] env[66641]: value = "task-5146093" [ 823.883187] env[66641]: _type = "Task" [ 823.883187] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.897857] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146093, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.934145] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 823.934145] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffb4b7dc-c0fc-4670-b319-0a8b48eace11 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.943132] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 823.943972] env[66641]: ERROR oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk due to incomplete transfer. [ 823.947228] env[66641]: DEBUG nova.compute.utils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 823.949713] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0c247900-cf35-44e1-9d15-db40e58468fc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.956116] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 823.956933] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 823.956933] env[66641]: WARNING neutronclient.v2_0.client [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.957101] env[66641]: WARNING neutronclient.v2_0.client [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 823.957799] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 823.959272] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 823.976494] env[66641]: DEBUG oslo_vmware.rw_handles [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/523ed499-a1b7-6217-c0ad-ff188989850e/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 823.976721] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Uploaded image 502c337b-699c-4132-a047-33d44a1bd1bd to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 823.980015] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 823.980015] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-8483275e-28dc-41f4-9db8-91031178a96b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.990708] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 823.990708] env[66641]: value = "task-5146094" [ 823.990708] env[66641]: _type = "Task" [ 823.990708] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.010293] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146094, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.019274] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updated VIF entry in instance network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 824.019274] env[66641]: DEBUG nova.network.neutron [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 824.132952] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146092, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.178789] env[66641]: DEBUG nova.policy [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '054d6ab93d3c4bcfa0ed71e59235751b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3563c790f3d745dc8dd34432a0b2fa5c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 824.264376] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bd92a3d-c1dc-4cb3-bff2-a0aa673ebe52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.274643] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f45c0c-ff81-418b-a596-f4148ac20ba6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.316081] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c48e3a1-6d66-4f18-9dcb-caaac6af43ee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.326448] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58279556-2f7f-4ff8-a538-3134d20f8d4a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.342876] env[66641]: DEBUG nova.compute.provider_tree [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 824.358566] env[66641]: DEBUG oslo_vmware.api [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146091, 'name': PowerOnVM_Task, 'duration_secs': 0.689982} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.358566] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 824.358566] env[66641]: INFO nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Took 10.01 seconds to spawn the instance on the hypervisor. [ 824.358566] env[66641]: DEBUG nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 824.358566] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a57810d7-1f17-4fba-b57d-fd90c1dd2c9a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.395096] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146093, 'name': ReconfigVM_Task, 'duration_secs': 0.250083} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.395416] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfigured VM instance instance-0000001e to detach disk 2000 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 824.397128] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b492ff-d77c-483a-82e6-805c994f1789 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.428243] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfiguring VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 824.428921] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db7be5ca-960c-4b16-a772-28874c1ef6ef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.450621] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 824.450621] env[66641]: value = "task-5146095" [ 824.450621] env[66641]: _type = "Task" [ 824.450621] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.455428] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 824.465115] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146095, 'name': ReconfigVM_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.502183] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146094, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.521665] env[66641]: DEBUG oslo_concurrency.lockutils [req-665e5f83-2bfd-4c15-b84a-52d133fd30d4 req-7d663d06-784e-4e5d-b4f4-5fa044668eca service nova] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 824.626235] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146092, 'name': CreateSnapshot_Task, 'duration_secs': 0.982142} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.626567] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 824.627350] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819954cc-1862-4b34-b516-23102590da6c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.633863] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Successfully created port: 8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 824.850816] env[66641]: DEBUG nova.scheduler.client.report [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 824.890171] env[66641]: INFO nova.compute.manager [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Took 31.04 seconds to build instance. [ 824.969446] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146095, 'name': ReconfigVM_Task, 'duration_secs': 0.413533} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.969745] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Reconfigured VM instance instance-0000001e to attach disk [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7/58fefaa4-0b17-408f-9329-78f8b5cf3fa7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.971962] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14331dd-2422-4bcd-afc1-9dd2c47072c2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.992991] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30533316-2820-4980-aa3d-00b2bf614aa1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.020956] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146094, 'name': Destroy_Task, 'duration_secs': 0.732839} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.021411] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Destroyed the VM [ 825.021588] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 825.022367] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5c1b91-8c35-4674-b6bf-0b5872f6aa6a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.025144] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1bc9c44b-8fc2-4930-a115-545a7fbd623c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.048337] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c304cda3-9fb8-4186-b69c-ce7a59505dc3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.051976] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 825.051976] env[66641]: value = "task-5146096" [ 825.051976] env[66641]: _type = "Task" [ 825.051976] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.063404] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 825.064378] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8dbffb78-7d8d-4e39-a5eb-c163710f3580 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.070961] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146096, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.072534] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 825.072534] env[66641]: value = "task-5146097" [ 825.072534] env[66641]: _type = "Task" [ 825.072534] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.081862] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.151556] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 825.151985] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-15c172aa-b894-41f0-bed9-fb009f464e4a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.162464] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 825.162464] env[66641]: value = "task-5146098" [ 825.162464] env[66641]: _type = "Task" [ 825.162464] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.175157] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146098, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.394340] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c981a5e-43a0-4fec-9523-7abd2df53e65 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.550s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.475922] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 825.507183] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 825.507435] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 825.507593] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 825.507775] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 825.507919] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 825.508078] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 825.508302] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.508464] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 825.508630] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 825.508792] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 825.508965] env[66641]: DEBUG nova.virt.hardware [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 825.510247] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d715aa5a-7dbc-4755-aa5d-1ca045c84854 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.519282] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3776aeef-365b-4b94-a9d8-d79f433c76b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.562490] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146096, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.584031] env[66641]: DEBUG oslo_vmware.api [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146097, 'name': PowerOnVM_Task, 'duration_secs': 0.421184} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.584339] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 825.673780] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146098, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.861983] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.419s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.865052] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 8.151s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.865257] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 825.865451] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 825.865829] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.195s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 825.867341] env[66641]: INFO nova.compute.claims [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 825.870531] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7887574-897e-4f1c-935a-484f3cf7a42d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.879782] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f4fcf5-5059-49e5-9666-c3f4b58914fc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.895985] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d93a377-a87f-41a6-8d37-1842a08fdcad {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.904339] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dfee826-e16e-44ef-a59d-7f3a551b109c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.945499] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179006MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 825.945709] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 826.064778] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146096, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.180943] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146098, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.360781] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Successfully updated port: 8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 826.453864] env[66641]: INFO nova.scheduler.client.report [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleted allocation for migration e82278e1-5e69-4b99-8727-b88245abb346 [ 826.567085] env[66641]: DEBUG oslo_vmware.api [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146096, 'name': RemoveSnapshot_Task, 'duration_secs': 1.071135} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.567367] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 826.567639] env[66641]: INFO nova.compute.manager [None req-3f17bd7b-f1b0-4a7e-aa88-051485717917 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 18.37 seconds to snapshot the instance on the hypervisor. [ 826.598556] env[66641]: INFO nova.compute.manager [None req-835d3a52-8d6d-433e-baef-49b1bf19dbbe tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance to original state: 'active' [ 826.677038] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146098, 'name': CloneVM_Task, 'duration_secs': 1.419891} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.677775] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Created linked-clone VM from snapshot [ 826.678295] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7010f35b-8e12-4f61-b38a-888722a0f910 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.688289] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Uploading image fa36f4ca-cd2e-4a6e-b766-2a464df3b3cc {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 826.723030] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 826.723030] env[66641]: value = "vm-1000703" [ 826.723030] env[66641]: _type = "VirtualMachine" [ 826.723030] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 826.723030] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-92ad9cf0-167f-4025-980d-aa4dec18e6d5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.731414] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lease: (returnval){ [ 826.731414] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eb87b9-9277-cd38-2e0b-baf7167bc092" [ 826.731414] env[66641]: _type = "HttpNfcLease" [ 826.731414] env[66641]: } obtained for exporting VM: (result){ [ 826.731414] env[66641]: value = "vm-1000703" [ 826.731414] env[66641]: _type = "VirtualMachine" [ 826.731414] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 826.732133] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the lease: (returnval){ [ 826.732133] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eb87b9-9277-cd38-2e0b-baf7167bc092" [ 826.732133] env[66641]: _type = "HttpNfcLease" [ 826.732133] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 826.739500] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 826.739500] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eb87b9-9277-cd38-2e0b-baf7167bc092" [ 826.739500] env[66641]: _type = "HttpNfcLease" [ 826.739500] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 826.864140] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.865025] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquired lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 826.865025] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 826.963718] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 16.379s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.131149] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-439ab40e-c014-40dc-b109-162e18ae94cb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.143787] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58bc1456-efa7-40ce-8d86-37bdaa61d68b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.184493] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8ab087-a0fb-4007-8d04-f18bfba2c5aa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.194699] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52d9517a-f43b-4221-a178-0a164c3edb04 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.217495] env[66641]: DEBUG nova.compute.provider_tree [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.244197] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 827.244197] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eb87b9-9277-cd38-2e0b-baf7167bc092" [ 827.244197] env[66641]: _type = "HttpNfcLease" [ 827.244197] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 827.244197] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 827.244197] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52eb87b9-9277-cd38-2e0b-baf7167bc092" [ 827.244197] env[66641]: _type = "HttpNfcLease" [ 827.244197] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 827.244197] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9981af85-cd6c-4d9e-bc5d-c16c8eb5efbb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.252889] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 827.253843] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 827.368581] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.368869] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.374684] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 827.519977] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f6d3bd47-aba6-4533-b507-e7689eb5c9bb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.712664] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.715082] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.723090] env[66641]: DEBUG nova.scheduler.client.report [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 827.841662] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 827.842093] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 827.856158] env[66641]: DEBUG nova.compute.manager [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Received event network-vif-plugged-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 827.857192] env[66641]: DEBUG oslo_concurrency.lockutils [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] Acquiring lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 827.857515] env[66641]: DEBUG oslo_concurrency.lockutils [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 827.857745] env[66641]: DEBUG oslo_concurrency.lockutils [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 827.857951] env[66641]: DEBUG nova.compute.manager [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] No waiting events found dispatching network-vif-plugged-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 827.858846] env[66641]: WARNING nova.compute.manager [req-607f3e30-5161-4015-aad4-849177c4b1f7 req-89a18a6a-c0ee-474a-81da-472ac8eadc57 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Received unexpected event network-vif-plugged-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 for instance with vm_state building and task_state spawning. [ 828.017848] env[66641]: DEBUG nova.network.neutron [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Updating instance_info_cache with network_info: [{"id": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "address": "fa:16:3e:c4:e1:22", "network": {"id": "2293d60d-2c06-4e84-8a53-7f70077f2952", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-624890248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3563c790f3d745dc8dd34432a0b2fa5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b30e411-0e", "ovs_interfaceid": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 828.129828] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.130305] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.130978] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 828.131317] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.131615] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.133942] env[66641]: INFO nova.compute.manager [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Terminating instance [ 828.229934] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 828.230168] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 828.236290] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.743s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 828.239681] env[66641]: INFO nova.compute.claims [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.526583] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Releasing lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 828.526583] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Instance network_info: |[{"id": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "address": "fa:16:3e:c4:e1:22", "network": {"id": "2293d60d-2c06-4e84-8a53-7f70077f2952", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-624890248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3563c790f3d745dc8dd34432a0b2fa5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b30e411-0e", "ovs_interfaceid": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 828.527032] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c4:e1:22', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ea4a9e02-45f1-4afb-8abb-0de26b153086', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.538055] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Creating folder: Project (3563c790f3d745dc8dd34432a0b2fa5c). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.538499] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8953732-8fb1-4dab-95d9-a205836dab1c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.551836] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Created folder: Project (3563c790f3d745dc8dd34432a0b2fa5c) in parent group-v1000566. [ 828.552151] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Creating folder: Instances. Parent ref: group-v1000704. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.552782] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9881444e-32b8-49bc-b7f3-d3a9b4754b06 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.565016] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Created folder: Instances in parent group-v1000704. [ 828.565308] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 828.565525] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.565741] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6041c39d-1b30-4975-bbd0-794c541a9a37 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.594159] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.594159] env[66641]: value = "task-5146102" [ 828.594159] env[66641]: _type = "Task" [ 828.594159] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.608106] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146102, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.641423] env[66641]: DEBUG nova.compute.manager [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 828.642399] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 828.644324] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f30337-2b48-475d-8457-ed5694901c4b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.658792] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 828.659719] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d8504b74-a9ec-417c-b325-d2c9ce612c7b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.668917] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 828.668917] env[66641]: value = "task-5146103" [ 828.668917] env[66641]: _type = "Task" [ 828.668917] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.685566] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146103, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.726743] env[66641]: DEBUG nova.compute.manager [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 828.727399] env[66641]: DEBUG nova.compute.manager [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing instance network info cache due to event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 828.727811] env[66641]: DEBUG oslo_concurrency.lockutils [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Acquiring lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.728206] env[66641]: DEBUG oslo_concurrency.lockutils [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Acquired lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 828.728708] env[66641]: DEBUG nova.network.neutron [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 828.745838] env[66641]: DEBUG nova.compute.utils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 828.751851] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 829.106120] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146102, 'name': CreateVM_Task, 'duration_secs': 0.481365} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.106432] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 829.107275] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.108149] env[66641]: WARNING openstack [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.120795] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.120919] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.121352] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 829.121694] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-25296a43-8e93-49fa-8182-f56e9f4b9352 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.131873] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 829.131873] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b48c33-32c4-75a0-9d8f-6ff68831b83b" [ 829.131873] env[66641]: _type = "Task" [ 829.131873] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.159593] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b48c33-32c4-75a0-9d8f-6ff68831b83b, 'name': SearchDatastore_Task, 'duration_secs': 0.012578} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.160702] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.161153] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.161535] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.161708] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquired lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 829.161912] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.162526] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b8bd5752-9bdb-42c4-a7ba-8981732be48d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.180890] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.180890] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.184399] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7138024e-3d52-4cb9-82e2-5f1710535531 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.193670] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 829.193670] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524eedea-85e5-bf51-1f71-96c0c9804fc3" [ 829.193670] env[66641]: _type = "Task" [ 829.193670] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.204307] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146103, 'name': PowerOffVM_Task, 'duration_secs': 0.281081} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.213022] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 829.213022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 829.213022] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8683aea-51eb-4f0d-8b32-6c667b309ab1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.224025] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524eedea-85e5-bf51-1f71-96c0c9804fc3, 'name': SearchDatastore_Task, 'duration_secs': 0.014928} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.224025] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26d776eb-57d9-4fc4-8ff1-fa9ea1986a93 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.233029] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.233501] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 829.241464] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 829.241464] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d8c875-fae8-5410-33bf-99e9a5a33dfb" [ 829.241464] env[66641]: _type = "Task" [ 829.241464] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.254350] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 829.263110] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d8c875-fae8-5410-33bf-99e9a5a33dfb, 'name': SearchDatastore_Task, 'duration_secs': 0.019366} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.264279] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Releasing lock "[datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 829.265033] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] acde41f9-4256-4e63-98e3-fd092c66c71e/acde41f9-4256-4e63-98e3-fd092c66c71e.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 829.265833] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8a10f0a7-5c4a-4395-8a66-6994b6a96163 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.280220] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 829.280220] env[66641]: value = "task-5146105" [ 829.280220] env[66641]: _type = "Task" [ 829.280220] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.289949] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146105, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.358280] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 829.358513] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 829.358874] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleting the datastore file [datastore2] 48f99287-b737-45fa-ad59-9e1425afa3d5 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.359639] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6386b247-ce8e-4d30-8a6f-c1615d31a660 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.379094] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for the task: (returnval){ [ 829.379094] env[66641]: value = "task-5146106" [ 829.379094] env[66641]: _type = "Task" [ 829.379094] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.395272] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146106, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.568050] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672f8ab9-e267-42bb-b3cf-c422039b1396 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.579724] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5193ab-3d78-4a9d-bd80-c1d0107b8704 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.625466] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9232838-0df1-406d-b9be-0f5a4387957c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.635023] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26be13c3-4ebf-4f32-be35-0743b6a7aa5f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.642364] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.642718] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.643015] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 829.643277] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 829.643551] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 829.658672] env[66641]: DEBUG nova.compute.provider_tree [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.664250] env[66641]: INFO nova.compute.manager [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Terminating instance [ 829.795110] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146105, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.890447] env[66641]: DEBUG oslo_vmware.api [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Task: {'id': task-5146106, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.341363} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.890689] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.890890] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 829.891139] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.891397] env[66641]: INFO nova.compute.manager [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Took 1.25 seconds to destroy the instance on the hypervisor. [ 829.891747] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 829.892313] env[66641]: DEBUG nova.compute.manager [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 829.892440] env[66641]: DEBUG nova.network.neutron [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 829.893513] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 829.893829] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.104996] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.110208] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.165336] env[66641]: DEBUG nova.scheduler.client.report [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 830.171194] env[66641]: DEBUG nova.compute.manager [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 830.171446] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 830.172574] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2357d3a2-70f4-4825-8ecb-66a17323e739 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.182815] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 830.183361] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-82a9ef8a-db09-4e06-a3d1-61cc2036956a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.192033] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 830.192033] env[66641]: value = "task-5146107" [ 830.192033] env[66641]: _type = "Task" [ 830.192033] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.204790] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146107, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.213596] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.213894] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.273591] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 830.303485] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146105, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.552379} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.307727] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 830.308053] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 830.308542] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 830.308783] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 830.308933] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 830.309105] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 830.309641] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.309641] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 830.309641] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 830.309823] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 830.309928] env[66641]: DEBUG nova.virt.hardware [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 830.310357] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore1] acde41f9-4256-4e63-98e3-fd092c66c71e/acde41f9-4256-4e63-98e3-fd092c66c71e.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 830.310557] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 830.311556] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0325a7e7-1393-43d6-af16-5fb8e275c620 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.314880] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-264bd52d-6e11-49da-aa7e-d843aeddd42b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.325275] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ee40a94-6217-49f6-aeef-d6b4709792ec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.330910] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 830.330910] env[66641]: value = "task-5146108" [ 830.330910] env[66641]: _type = "Task" [ 830.330910] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.348317] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 830.354693] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Creating folder: Project (a1f76c8110d542c78ff65bec969178c5). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.356788] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac01f2c8-5dd9-423b-98f3-08cd97d3d41d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.360566] env[66641]: DEBUG nova.compute.manager [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Received event network-changed-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 830.361024] env[66641]: DEBUG nova.compute.manager [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Refreshing instance network info cache due to event network-changed-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 830.361450] env[66641]: DEBUG oslo_concurrency.lockutils [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Acquiring lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.361620] env[66641]: DEBUG oslo_concurrency.lockutils [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Acquired lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 830.361816] env[66641]: DEBUG nova.network.neutron [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Refreshing network info cache for port 8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 830.367137] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146108, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.380915] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Created folder: Project (a1f76c8110d542c78ff65bec969178c5) in parent group-v1000566. [ 830.381297] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Creating folder: Instances. Parent ref: group-v1000707. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 830.381478] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7eb09248-eca1-48ee-a6de-56a2ee627d23 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.395441] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Created folder: Instances in parent group-v1000707. [ 830.395755] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 830.396029] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 830.396414] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-65ead1ef-cecf-4723-8fb4-76131477ddc9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.416792] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 830.416792] env[66641]: value = "task-5146111" [ 830.416792] env[66641]: _type = "Task" [ 830.416792] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.427461] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146111, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.673544] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 830.674255] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 830.678034] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.170s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 830.680390] env[66641]: INFO nova.compute.claims [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.704899] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146107, 'name': PowerOffVM_Task, 'duration_secs': 0.256978} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.705233] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 830.705417] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 830.706584] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ecbf14e-8ced-41dc-8776-b88e684f1196 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.782558] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 830.782788] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 830.782968] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleting the datastore file [datastore2] 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 830.783371] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b0d58c99-9bb3-4454-8bc4-55e12c3d79f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.792409] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 830.792409] env[66641]: value = "task-5146113" [ 830.792409] env[66641]: _type = "Task" [ 830.792409] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.804356] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146113, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.841479] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146108, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085771} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.842041] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 830.842750] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6302006-7569-4125-9001-28ff92ae0e9f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.866702] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Reconfiguring VM instance instance-0000002d to attach disk [datastore1] acde41f9-4256-4e63-98e3-fd092c66c71e/acde41f9-4256-4e63-98e3-fd092c66c71e.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 830.867047] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2bb2e328-5560-47b2-8244-1d00ec814efa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.883693] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 830.884215] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 830.897604] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 830.897604] env[66641]: value = "task-5146114" [ 830.897604] env[66641]: _type = "Task" [ 830.897604] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.906232] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146114, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.927770] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146111, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.053582] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.054331] env[66641]: WARNING openstack [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.067458] env[66641]: DEBUG nova.network.neutron [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 831.190432] env[66641]: DEBUG nova.compute.utils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 831.198028] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 831.198028] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 831.198028] env[66641]: WARNING neutronclient.v2_0.client [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.198028] env[66641]: WARNING neutronclient.v2_0.client [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 831.198028] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.198348] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.314127] env[66641]: DEBUG nova.policy [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3a37fbe85d6446d5a35bf17064fabd96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '26e21ce2b8bc4f209cc4ac5453d151e7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 831.329035] env[66641]: DEBUG oslo_vmware.api [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146113, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.529226} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.330416] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 831.330638] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 831.330856] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 831.331117] env[66641]: INFO nova.compute.manager [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 831.331394] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 831.331625] env[66641]: DEBUG nova.compute.manager [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 831.331712] env[66641]: DEBUG nova.network.neutron [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 831.332267] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.332520] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.382516] env[66641]: DEBUG nova.network.neutron [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updated VIF entry in instance network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 831.382867] env[66641]: DEBUG nova.network.neutron [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [{"id": "049d8c0f-cda0-4cd9-b972-188efe21a140", "address": "fa:16:3e:15:f5:a4", "network": {"id": "28059287-894a-40a1-8385-d1fddec6f422", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-167219310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a9466d8e6fa49c59a14110b417664c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap049d8c0f-cd", "ovs_interfaceid": "049d8c0f-cda0-4cd9-b972-188efe21a140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 831.414215] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146114, 'name': ReconfigVM_Task, 'duration_secs': 0.455546} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.416574] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Reconfigured VM instance instance-0000002d to attach disk [datastore1] acde41f9-4256-4e63-98e3-fd092c66c71e/acde41f9-4256-4e63-98e3-fd092c66c71e.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 831.419032] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed3c8b1f-4afd-46b8-b9c8-8726d7b09029 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.429194] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.429574] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.445128] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 831.445128] env[66641]: value = "task-5146115" [ 831.445128] env[66641]: _type = "Task" [ 831.445128] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.445726] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146111, 'name': CreateVM_Task, 'duration_secs': 0.692304} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.446240] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 831.450192] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 831.450192] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 831.450320] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 831.450617] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74c03720-5099-4760-90b5-67e3b62f563a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.461771] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146115, 'name': Rename_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.463852] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 831.463852] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d68f16-4ff6-0e53-dd19-ba746ea2ba30" [ 831.463852] env[66641]: _type = "Task" [ 831.463852] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.475638] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d68f16-4ff6-0e53-dd19-ba746ea2ba30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.575392] env[66641]: INFO nova.compute.manager [-] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Took 1.68 seconds to deallocate network for instance. [ 831.580360] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.581537] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 831.623879] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40875251-8c09-4f85-a495-a7d2a72dfab3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.640274] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c677bd-46a4-44ac-bafd-955e26dc4be3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.686140] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd972b4-087b-4f03-bccb-0f559f286aa9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.695310] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056823d7-191a-4613-9a3d-b552ecf31d23 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.705585] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 831.720133] env[66641]: DEBUG nova.compute.provider_tree [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.888832] env[66641]: DEBUG oslo_concurrency.lockutils [req-2722a646-230b-49a1-9675-6da44b48efe5 req-c36afeec-a28f-4486-aa8a-005d5d1d0adf service nova] Releasing lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 831.932422] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Successfully created port: a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 831.958337] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146115, 'name': Rename_Task, 'duration_secs': 0.27713} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.959047] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 831.961738] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4df50e5c-9123-4c5b-ad8c-f5fcf27b95c6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.972917] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 831.972917] env[66641]: value = "task-5146116" [ 831.972917] env[66641]: _type = "Task" [ 831.972917] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.986746] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 831.986746] env[66641]: WARNING openstack [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 832.000053] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52d68f16-4ff6-0e53-dd19-ba746ea2ba30, 'name': SearchDatastore_Task, 'duration_secs': 0.024569} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.001034] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 832.001327] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 832.001588] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.001704] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 832.001962] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 832.002762] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a0820cab-cbbd-4be8-9d1f-a2e3de981b67 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.014138] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146116, 'name': PowerOnVM_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.020410] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 832.021056] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 832.021750] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0317e1a6-1234-452b-9dd0-869db0339773 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.029856] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 832.029856] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525992d5-1dde-716d-f872-9e9cd26a6a69" [ 832.029856] env[66641]: _type = "Task" [ 832.029856] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.039735] env[66641]: DEBUG nova.compute.manager [req-3f346bcf-0439-48b7-9cc9-f1470fea785a req-9d2c0f33-d0a2-48b3-9262-be013ee75d65 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Received event network-vif-deleted-a44fab6d-1a8a-4a32-93c1-ebfd24e3d021 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 832.039956] env[66641]: INFO nova.compute.manager [req-3f346bcf-0439-48b7-9cc9-f1470fea785a req-9d2c0f33-d0a2-48b3-9262-be013ee75d65 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Neutron deleted interface a44fab6d-1a8a-4a32-93c1-ebfd24e3d021; detaching it from the instance and deleting it from the info cache [ 832.040179] env[66641]: DEBUG nova.network.neutron [req-3f346bcf-0439-48b7-9cc9-f1470fea785a req-9d2c0f33-d0a2-48b3-9262-be013ee75d65 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.045830] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525992d5-1dde-716d-f872-9e9cd26a6a69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.109275] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 832.225733] env[66641]: DEBUG nova.scheduler.client.report [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 832.464670] env[66641]: DEBUG nova.network.neutron [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 832.492908] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146116, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.541672] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]525992d5-1dde-716d-f872-9e9cd26a6a69, 'name': SearchDatastore_Task, 'duration_secs': 0.01944} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.542654] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7be9c12-a6a7-4112-aa73-796f4116ea30 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.547559] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a5844df0-3609-46ba-ab55-f649defda224 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.551782] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 832.551782] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a49b37-6e81-c4f2-b844-914baf25bf28" [ 832.551782] env[66641]: _type = "Task" [ 832.551782] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.562950] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52da49a0-f5fa-41d6-aa73-75af9ec0a905 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.579509] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a49b37-6e81-c4f2-b844-914baf25bf28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.603096] env[66641]: DEBUG nova.compute.manager [req-3f346bcf-0439-48b7-9cc9-f1470fea785a req-9d2c0f33-d0a2-48b3-9262-be013ee75d65 service nova] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Detach interface failed, port_id=a44fab6d-1a8a-4a32-93c1-ebfd24e3d021, reason: Instance 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 832.717183] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 832.732450] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.054s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 832.733522] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 832.737163] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.791s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 832.758430] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 832.758430] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 832.758430] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 832.758625] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 832.759027] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 832.759027] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 832.759192] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.759408] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 832.759617] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 832.759789] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 832.759956] env[66641]: DEBUG nova.virt.hardware [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 832.761186] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78574014-abc7-44f2-b9b4-6016292686ac {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.770995] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b57284d-3d7b-4ed3-8adb-f3ee3d0c32bd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.968988] env[66641]: INFO nova.compute.manager [-] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Took 1.64 seconds to deallocate network for instance. [ 832.993196] env[66641]: DEBUG oslo_vmware.api [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146116, 'name': PowerOnVM_Task, 'duration_secs': 0.815238} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.993196] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 832.993196] env[66641]: INFO nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Took 7.52 seconds to spawn the instance on the hypervisor. [ 832.993833] env[66641]: DEBUG nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 832.994756] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5173c811-10e5-4ef7-9a9d-e62b62806759 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.030943] env[66641]: DEBUG nova.network.neutron [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Updated VIF entry in instance network info cache for port 8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 833.031335] env[66641]: DEBUG nova.network.neutron [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Updating instance_info_cache with network_info: [{"id": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "address": "fa:16:3e:c4:e1:22", "network": {"id": "2293d60d-2c06-4e84-8a53-7f70077f2952", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-624890248-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3563c790f3d745dc8dd34432a0b2fa5c", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ea4a9e02-45f1-4afb-8abb-0de26b153086", "external-id": "nsx-vlan-transportzone-336", "segmentation_id": 336, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8b30e411-0e", "ovs_interfaceid": "8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 833.071325] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52a49b37-6e81-c4f2-b844-914baf25bf28, 'name': SearchDatastore_Task, 'duration_secs': 0.017527} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.073051] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.073051] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 833.073051] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-acc5683f-8f12-43d2-b4c8-79d78a8ff954 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.083751] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 833.083751] env[66641]: value = "task-5146117" [ 833.083751] env[66641]: _type = "Task" [ 833.083751] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.102812] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146117, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.210903] env[66641]: DEBUG nova.compute.manager [req-ebcd13d3-b8d4-48a5-9bb3-b37910da5d5f req-df0d1372-dd7b-4ae9-8670-6ac38d2f6eef service nova] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Received event network-vif-deleted-cdd4e9c6-89a9-4a63-8850-26be458c02fe {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 833.241763] env[66641]: DEBUG nova.compute.utils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 833.252013] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 833.252527] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 833.252732] env[66641]: WARNING neutronclient.v2_0.client [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.253167] env[66641]: WARNING neutronclient.v2_0.client [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 833.253763] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 833.254163] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 833.481296] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 833.524383] env[66641]: INFO nova.compute.manager [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Took 22.35 seconds to build instance. [ 833.536884] env[66641]: DEBUG oslo_concurrency.lockutils [req-5173699d-8dc7-411b-89fb-e39fc0eceac4 req-8b7f884f-a3c7-4af9-9099-21eb917e1f27 service nova] Releasing lock "refresh_cache-acde41f9-4256-4e63-98e3-fd092c66c71e" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 833.600953] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146117, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.641103] env[66641]: DEBUG nova.policy [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ee8a189cf0404683981bf01b49928c32', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2cbdc2e8f9047d98c5cbaa1c409f4d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 833.718326] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Successfully updated port: a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 833.756314] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 833.797313] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797313] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance a7575399-f72d-4d2c-b57b-f2f2f591c8a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797313] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance e508b1e9-cbb0-408c-98e1-d54ceaec1f1b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797313] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797472] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 40d7571f-09e2-463a-a449-36c621045819 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797472] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance c95b481a-7956-410d-971c-7d94911230bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797472] env[66641]: WARNING nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 48f99287-b737-45fa-ad59-9e1425afa3d5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 833.797645] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance cdfcd9d8-dbf7-4046-8338-42762fc389a1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.797898] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 88ae00a2-6139-4258-b316-0f75032275ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.798070] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance acde41f9-4256-4e63-98e3-fd092c66c71e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.798320] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance a4f27205-f8ce-49f9-a3de-2e53bcbb6527 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.798481] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 4e70780a-62ca-4e4a-9366-00dc35750c61 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.798710] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 29f03150-f0a2-4424-b758-32891730923c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 833.798980] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 12 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 833.799277] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2880MB phys_disk=100GB used_disk=12GB total_vcpus=48 used_vcpus=12 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '12', 'num_vm_active': '8', 'num_task_None': '7', 'num_os_type_None': '12', 'num_proj_6624f269bb8840e98c10259f006ce962': '2', 'io_workload': '4', 'num_proj_7ecf362f0f454995adbf560fa17e1dab': '2', 'num_task_deleting': '1', 'num_task_image_uploading': '1', 'num_proj_17276c1118a6476b81390021f8fe9ed2': '1', 'num_proj_c48b059b6f7d49a8ba5ecf28125ddaa9': '1', 'num_proj_f7e0123f065d40fe8b51e997ae3f6089': '1', 'num_proj_3a9466d8e6fa49c59a14110b417664c6': '1', 'num_vm_building': '4', 'num_task_spawning': '3', 'num_proj_3563c790f3d745dc8dd34432a0b2fa5c': '1', 'num_proj_a1f76c8110d542c78ff65bec969178c5': '1', 'num_proj_26e21ce2b8bc4f209cc4ac5453d151e7': '1', 'num_proj_c2cbdc2e8f9047d98c5cbaa1c409f4d3': '1'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 834.030798] env[66641]: DEBUG oslo_concurrency.lockutils [None req-646dbebf-5652-438f-ad78-000b3350a084 tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.875s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.056746] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8c6a55c-a035-4186-ae77-62239833c8b7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.066064] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10058cdd-11ea-46b5-91d2-623ebe19f03b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.105537] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f269d51-3348-4368-b6c2-f3014ef9cbbe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.118097] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146117, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.646242} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.120948] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 834.121457] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 834.121953] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f6ec2ec-12ce-4bfa-942d-3a34d8f05f14 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.126211] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f597d7fa-60e7-4bf5-8a85-c75d378e5e34 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.146497] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.150256] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 834.150256] env[66641]: value = "task-5146118" [ 834.150256] env[66641]: _type = "Task" [ 834.150256] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.153075] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Successfully created port: 4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 834.168094] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146118, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.205818] env[66641]: DEBUG nova.compute.manager [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Received event network-vif-plugged-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 834.206834] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Acquiring lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 834.206834] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 834.206834] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 834.207039] env[66641]: DEBUG nova.compute.manager [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] No waiting events found dispatching network-vif-plugged-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 834.207136] env[66641]: WARNING nova.compute.manager [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Received unexpected event network-vif-plugged-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 for instance with vm_state building and task_state spawning. [ 834.207302] env[66641]: DEBUG nova.compute.manager [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Received event network-changed-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 834.207436] env[66641]: DEBUG nova.compute.manager [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Refreshing instance network info cache due to event network-changed-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 834.207644] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Acquiring lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.207776] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Acquired lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 834.207925] env[66641]: DEBUG nova.network.neutron [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Refreshing network info cache for port a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 834.224019] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.659693] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 834.674970] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146118, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077431} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.677441] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 834.680325] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f22c24e-6da6-4d4f-8938-f094b00e10ea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.707159] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 834.707853] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-dcd16dfa-8658-4ff8-8f73-0a85de05f776 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.724341] env[66641]: WARNING openstack [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 834.724762] env[66641]: WARNING openstack [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 834.729660] env[66641]: DEBUG nova.network.neutron [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 834.739677] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 834.739677] env[66641]: value = "task-5146119" [ 834.739677] env[66641]: _type = "Task" [ 834.739677] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.749856] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.764578] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 834.792316] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 834.792596] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 834.792769] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 834.792953] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 834.793665] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 834.793848] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 834.794240] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 834.794429] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 834.794614] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 834.794782] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 834.794966] env[66641]: DEBUG nova.virt.hardware [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 834.795892] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73950c26-19ca-4a02-a5db-0e5e9a6ea11c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.805908] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939e9bf8-f342-409f-972f-d688b70c6c0a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.162975] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 835.163388] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.426s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.163714] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.055s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.163909] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.166574] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.685s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.166574] env[66641]: DEBUG nova.objects.instance [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lazy-loading 'resources' on Instance uuid 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 835.189413] env[66641]: INFO nova.scheduler.client.report [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Deleted allocations for instance 48f99287-b737-45fa-ad59-9e1425afa3d5 [ 835.253034] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146119, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.459670] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "acde41f9-4256-4e63-98e3-fd092c66c71e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.459981] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.460211] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.460514] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.460630] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.463485] env[66641]: INFO nova.compute.manager [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Terminating instance [ 835.583767] env[66641]: DEBUG nova.network.neutron [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 835.585364] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.585628] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.585863] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.586066] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.586236] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.594049] env[66641]: INFO nova.compute.manager [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Terminating instance [ 835.679140] env[66641]: DEBUG nova.compute.manager [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Received event network-vif-plugged-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 835.679395] env[66641]: DEBUG oslo_concurrency.lockutils [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] Acquiring lock "29f03150-f0a2-4424-b758-32891730923c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 835.680072] env[66641]: DEBUG oslo_concurrency.lockutils [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] Lock "29f03150-f0a2-4424-b758-32891730923c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 835.680282] env[66641]: DEBUG oslo_concurrency.lockutils [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] Lock "29f03150-f0a2-4424-b758-32891730923c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.680458] env[66641]: DEBUG nova.compute.manager [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] No waiting events found dispatching network-vif-plugged-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 835.680622] env[66641]: WARNING nova.compute.manager [req-609afae0-f849-4197-b144-af7555e8558c req-f7a94b6d-ac5d-4738-a83b-c7656af745df service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Received unexpected event network-vif-plugged-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 for instance with vm_state building and task_state spawning. [ 835.702830] env[66641]: DEBUG oslo_concurrency.lockutils [None req-df052e34-001b-4881-80b0-649d5454cf67 tempest-DeleteServersTestJSON-1828863045 tempest-DeleteServersTestJSON-1828863045-project-member] Lock "48f99287-b737-45fa-ad59-9e1425afa3d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.572s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 835.754212] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146119, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.829646] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Successfully updated port: 4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 835.924637] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af516648-5c5d-4313-a067-2bc7ebe35f3c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.934511] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b8aae8c-a31d-46bf-89dc-a1ae7245b81b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.986952] env[66641]: DEBUG nova.compute.manager [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 835.987969] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.990673] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-526657ce-ae5e-4bd5-9a04-f7edfebd8ebc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.997313] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a58302e-326e-43e9-a324-7199da7cbe59 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.011355] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f332e7bb-3930-4667-9293-5dd37a26afd5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.017456] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.017993] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f964ffa-52fd-48ad-829b-6019db2e4a5b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.035768] env[66641]: DEBUG nova.compute.provider_tree [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.039501] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 836.039501] env[66641]: value = "task-5146120" [ 836.039501] env[66641]: _type = "Task" [ 836.039501] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.049754] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146120, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.092228] env[66641]: DEBUG oslo_concurrency.lockutils [req-8841aaa3-50ec-4a22-99ee-96e4f71e6071 req-d995d57f-71e3-4175-91d8-4dcb4023e591 service nova] Releasing lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 836.092806] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquired lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.093433] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 836.105315] env[66641]: DEBUG nova.compute.manager [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 836.105315] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 836.106569] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72dd7bad-f908-4bd4-9820-7f24515cb22a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.116464] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 836.116778] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b34efe2-04f5-4236-9c13-d410606fdc63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.125772] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 836.125772] env[66641]: value = "task-5146121" [ 836.125772] env[66641]: _type = "Task" [ 836.125772] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.137436] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146121, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.255711] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146119, 'name': ReconfigVM_Task, 'duration_secs': 1.441856} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.256156] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Reconfigured VM instance instance-0000002e to attach disk [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 836.257119] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3b1acbf6-cf34-4b6b-9bdf-7fdca5df9128 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.266401] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 836.266401] env[66641]: value = "task-5146122" [ 836.266401] env[66641]: _type = "Task" [ 836.266401] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.279521] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146122, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.332747] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.333062] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquired lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 836.333284] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 836.543032] env[66641]: DEBUG nova.scheduler.client.report [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 836.558578] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146120, 'name': PowerOffVM_Task, 'duration_secs': 0.267301} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.558578] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.558578] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 836.558578] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66fee662-5b66-480c-87b7-25fba2760fc9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.601343] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.601343] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.606061] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 836.640443] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 836.640443] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 836.640443] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Deleting the datastore file [datastore1] acde41f9-4256-4e63-98e3-fd092c66c71e {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.640443] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5522db6-0ef3-41f6-adcb-d4e3782c5bea {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.650367] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146121, 'name': PowerOffVM_Task, 'duration_secs': 0.252789} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.651131] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 836.651318] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 836.651694] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-27958be3-3481-45a2-a084-07a4a6d436fc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.656157] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for the task: (returnval){ [ 836.656157] env[66641]: value = "task-5146124" [ 836.656157] env[66641]: _type = "Task" [ 836.656157] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.666314] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146124, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.709418] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.709961] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.730301] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 836.731098] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 836.731098] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleting the datastore file [datastore2] e508b1e9-cbb0-408c-98e1-d54ceaec1f1b {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 836.731098] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8e86865b-7852-4c50-9dd5-4db3d861ed26 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.740373] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 836.740373] env[66641]: value = "task-5146126" [ 836.740373] env[66641]: _type = "Task" [ 836.740373] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.751145] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146126, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.782763] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146122, 'name': Rename_Task, 'duration_secs': 0.229293} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.783093] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 836.783452] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f623da1-6cc8-42fd-ae78-e9781a2c70cb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.792051] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 836.792051] env[66641]: value = "task-5146127" [ 836.792051] env[66641]: _type = "Task" [ 836.792051] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.801564] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146127, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.834497] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.835365] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.847923] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.848405] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 836.854156] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 836.882914] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 836.884083] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76899f9-a72e-4085-8a20-686eb7ebaed2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.891973] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 836.892247] env[66641]: ERROR oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk due to incomplete transfer. [ 836.892475] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4498f495-c66f-4029-a012-902d88d0c07e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.907548] env[66641]: DEBUG oslo_vmware.rw_handles [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524dc14a-9a33-fe87-5ae0-e59f954971cf/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 836.907548] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Uploaded image fa36f4ca-cd2e-4a6e-b766-2a464df3b3cc to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 836.909740] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 836.910050] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-3ce84909-5ef2-4108-92bd-65f18239aa3e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.917881] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 836.917881] env[66641]: value = "task-5146128" [ 836.917881] env[66641]: _type = "Task" [ 836.917881] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.932523] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146128, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.977349] env[66641]: DEBUG nova.network.neutron [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Updating instance_info_cache with network_info: [{"id": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "address": "fa:16:3e:fc:7c:97", "network": {"id": "278f3697-035d-48dc-9175-7202147bad69", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1403293202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26e21ce2b8bc4f209cc4ac5453d151e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa48fa631-89", "ovs_interfaceid": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 836.988986] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 836.989400] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.051640] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.885s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.059828] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.060251] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.082189] env[66641]: INFO nova.scheduler.client.report [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted allocations for instance 58fefaa4-0b17-408f-9329-78f8b5cf3fa7 [ 837.147645] env[66641]: DEBUG nova.network.neutron [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Updating instance_info_cache with network_info: [{"id": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "address": "fa:16:3e:de:b6:40", "network": {"id": "d28f29b2-3153-486d-970d-940e514ce23d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-625702414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2cbdc2e8f9047d98c5cbaa1c409f4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4100a6ee-e7", "ovs_interfaceid": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.168904] env[66641]: DEBUG oslo_vmware.api [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Task: {'id': task-5146124, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.228273} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.170187] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.170535] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 837.170722] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.172612] env[66641]: INFO nova.compute.manager [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Took 1.18 seconds to destroy the instance on the hypervisor. [ 837.172612] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.172612] env[66641]: DEBUG nova.compute.manager [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 837.172612] env[66641]: DEBUG nova.network.neutron [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 837.172612] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.172983] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.217035] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.217418] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.250626] env[66641]: DEBUG oslo_vmware.api [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146126, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208105} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.250949] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 837.251160] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 837.251368] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 837.251578] env[66641]: INFO nova.compute.manager [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Took 1.15 seconds to destroy the instance on the hypervisor. [ 837.251892] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.252144] env[66641]: DEBUG nova.compute.manager [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 837.252330] env[66641]: DEBUG nova.network.neutron [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 837.252899] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.253188] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.311881] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146127, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.432586] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146128, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.467606] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 837.467606] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 837.483390] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Releasing lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.483899] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Instance network_info: |[{"id": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "address": "fa:16:3e:fc:7c:97", "network": {"id": "278f3697-035d-48dc-9175-7202147bad69", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1403293202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26e21ce2b8bc4f209cc4ac5453d151e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa48fa631-89", "ovs_interfaceid": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 837.485206] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:7c:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd69a4b11-8d65-435f-94a5-28f74a39a718', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a48fa631-89c4-45c7-bb90-dce5ffb9e9c6', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.492467] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Creating folder: Project (26e21ce2b8bc4f209cc4ac5453d151e7). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.493100] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1dc9c15f-f744-4c66-b309-7afbd72ed5cf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.507814] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Created folder: Project (26e21ce2b8bc4f209cc4ac5453d151e7) in parent group-v1000566. [ 837.508688] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Creating folder: Instances. Parent ref: group-v1000710. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.508688] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-23cdc97d-33ad-4720-9ad2-e86597bbde6d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.524868] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Created folder: Instances in parent group-v1000710. [ 837.525168] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.527486] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.527486] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f3bfe8e9-5141-4d88-ac85-3cbc93a5f3f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.548716] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.548716] env[66641]: value = "task-5146131" [ 837.548716] env[66641]: _type = "Task" [ 837.548716] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.557014] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146131, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.594457] env[66641]: DEBUG oslo_concurrency.lockutils [None req-62b96a47-4f57-4755-8a7d-036055748932 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "58fefaa4-0b17-408f-9329-78f8b5cf3fa7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.950s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 837.652321] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Releasing lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 837.652321] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Instance network_info: |[{"id": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "address": "fa:16:3e:de:b6:40", "network": {"id": "d28f29b2-3153-486d-970d-940e514ce23d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-625702414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2cbdc2e8f9047d98c5cbaa1c409f4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4100a6ee-e7", "ovs_interfaceid": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 837.652472] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:b6:40', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76e60ff4-204c-4f48-bd0e-2d5fa0a812ef', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4100a6ee-e7e0-4d1f-afe1-05af072a5ce8', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.661032] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Creating folder: Project (c2cbdc2e8f9047d98c5cbaa1c409f4d3). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.661032] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7dff9309-1b7f-47d8-9d6b-10366411484d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.675169] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Created folder: Project (c2cbdc2e8f9047d98c5cbaa1c409f4d3) in parent group-v1000566. [ 837.675169] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Creating folder: Instances. Parent ref: group-v1000713. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.675364] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5995c7e5-80e8-4eb7-b1a7-8a8b3e93f8db {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.686461] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Created folder: Instances in parent group-v1000713. [ 837.686461] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 837.686461] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.686679] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b34f21d7-1b10-4f1f-9158-00d004197207 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.709339] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.709339] env[66641]: value = "task-5146134" [ 837.709339] env[66641]: _type = "Task" [ 837.709339] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.719090] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146134, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.736859] env[66641]: DEBUG nova.compute.manager [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Received event network-changed-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 837.738457] env[66641]: DEBUG nova.compute.manager [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Refreshing instance network info cache due to event network-changed-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 837.738905] env[66641]: DEBUG oslo_concurrency.lockutils [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Acquiring lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.738997] env[66641]: DEBUG oslo_concurrency.lockutils [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Acquired lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 837.739957] env[66641]: DEBUG nova.network.neutron [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Refreshing network info cache for port 4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 837.790054] env[66641]: DEBUG nova.compute.manager [req-2bbd3674-cabf-4c66-b78d-268a23d264d5 req-74b26855-0d20-473f-b3ac-397d468ad4d3 service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Received event network-vif-deleted-ee7db498-4256-4265-9b73-6b6b2afa057c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 837.790054] env[66641]: INFO nova.compute.manager [req-2bbd3674-cabf-4c66-b78d-268a23d264d5 req-74b26855-0d20-473f-b3ac-397d468ad4d3 service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Neutron deleted interface ee7db498-4256-4265-9b73-6b6b2afa057c; detaching it from the instance and deleting it from the info cache [ 837.790054] env[66641]: DEBUG nova.network.neutron [req-2bbd3674-cabf-4c66-b78d-268a23d264d5 req-74b26855-0d20-473f-b3ac-397d468ad4d3 service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.807795] env[66641]: DEBUG oslo_vmware.api [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146127, 'name': PowerOnVM_Task, 'duration_secs': 0.834136} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.808251] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 837.808575] env[66641]: INFO nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Took 7.53 seconds to spawn the instance on the hypervisor. [ 837.808863] env[66641]: DEBUG nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 837.810279] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb6bfe6-4676-4472-b5bb-fdf8d47f9e1d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.929657] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146128, 'name': Destroy_Task, 'duration_secs': 0.729267} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.929791] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Destroyed the VM [ 837.930461] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 837.930461] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e4fe63db-98e3-4960-b8ac-a7771b4d6092 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.938651] env[66641]: DEBUG nova.network.neutron [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 837.940247] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 837.940247] env[66641]: value = "task-5146135" [ 837.940247] env[66641]: _type = "Task" [ 837.940247] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.953243] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146135, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.059597] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146131, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.221856] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146134, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.245635] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.246058] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.260588] env[66641]: DEBUG nova.network.neutron [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 838.294298] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d2d0164-f6c9-41fe-bee8-9d43cdb9d55b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.307690] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59394524-2d6f-4164-852e-378bac01430c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.340923] env[66641]: DEBUG nova.compute.manager [req-2bbd3674-cabf-4c66-b78d-268a23d264d5 req-74b26855-0d20-473f-b3ac-397d468ad4d3 service nova] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Detach interface failed, port_id=ee7db498-4256-4265-9b73-6b6b2afa057c, reason: Instance e508b1e9-cbb0-408c-98e1-d54ceaec1f1b could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 838.341726] env[66641]: INFO nova.compute.manager [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Took 19.69 seconds to build instance. [ 838.387901] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.388548] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.441568] env[66641]: INFO nova.compute.manager [-] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Took 1.27 seconds to deallocate network for instance. [ 838.454885] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146135, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.471142] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.471613] env[66641]: WARNING openstack [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.563372] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146131, 'name': CreateVM_Task, 'duration_secs': 0.547622} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.563642] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.564707] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.565201] env[66641]: WARNING openstack [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.571507] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.571725] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 838.572117] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 838.574944] env[66641]: DEBUG nova.network.neutron [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Updated VIF entry in instance network info cache for port 4100a6ee-e7e0-4d1f-afe1-05af072a5ce8. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 838.575335] env[66641]: DEBUG nova.network.neutron [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Updating instance_info_cache with network_info: [{"id": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "address": "fa:16:3e:de:b6:40", "network": {"id": "d28f29b2-3153-486d-970d-940e514ce23d", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-625702414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2cbdc2e8f9047d98c5cbaa1c409f4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76e60ff4-204c-4f48-bd0e-2d5fa0a812ef", "external-id": "nsx-vlan-transportzone-854", "segmentation_id": 854, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4100a6ee-e7", "ovs_interfaceid": "4100a6ee-e7e0-4d1f-afe1-05af072a5ce8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 838.576852] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c457d534-4edb-4e47-be9e-0152b47fd6a4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.585288] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 838.585288] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ceabd7-7f6c-f77f-ce30-52a0e42197aa" [ 838.585288] env[66641]: _type = "Task" [ 838.585288] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.598443] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ceabd7-7f6c-f77f-ce30-52a0e42197aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.696077] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.696448] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.697308] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.698189] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.698503] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.701437] env[66641]: INFO nova.compute.manager [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Terminating instance [ 838.723416] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146134, 'name': CreateVM_Task, 'duration_secs': 0.671388} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.723815] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.724441] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 838.724835] env[66641]: WARNING openstack [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 838.733392] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.763430] env[66641]: INFO nova.compute.manager [-] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Took 1.51 seconds to deallocate network for instance. [ 838.844180] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6183fb08-e933-45c4-a82e-b2f401839289 tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.203s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 838.953493] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 838.953812] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 838.954074] env[66641]: DEBUG nova.objects.instance [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lazy-loading 'resources' on Instance uuid acde41f9-4256-4e63-98e3-fd092c66c71e {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 838.955183] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146135, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.080212] env[66641]: DEBUG oslo_concurrency.lockutils [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] Releasing lock "refresh_cache-29f03150-f0a2-4424-b758-32891730923c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.080495] env[66641]: DEBUG nova.compute.manager [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Received event network-vif-deleted-8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 839.080672] env[66641]: INFO nova.compute.manager [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Neutron deleted interface 8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1; detaching it from the instance and deleting it from the info cache [ 839.080849] env[66641]: DEBUG nova.network.neutron [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 839.097487] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ceabd7-7f6c-f77f-ce30-52a0e42197aa, 'name': SearchDatastore_Task, 'duration_secs': 0.014058} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.098454] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.098849] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.098950] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.099130] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.099311] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.099599] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.100206] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 839.101067] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e162ccdb-2802-4323-9ecf-5b31f59ecaf2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.103040] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1bb4ae37-f79e-4db1-b07c-fb3c8e7ff7c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.110483] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 839.110483] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52812730-c023-94b4-ae60-3d174662f5cd" [ 839.110483] env[66641]: _type = "Task" [ 839.110483] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.115726] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.116017] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.117194] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42845f6a-575f-48b4-a579-65e70412c233 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.124777] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52812730-c023-94b4-ae60-3d174662f5cd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.128839] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 839.128839] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52484c5a-ab0e-5395-c934-75652a54cfa4" [ 839.128839] env[66641]: _type = "Task" [ 839.128839] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.143693] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52484c5a-ab0e-5395-c934-75652a54cfa4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.209896] env[66641]: DEBUG nova.compute.manager [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 839.209896] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 839.209896] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6927926b-fa2d-4c59-9aa7-8405c65f1376 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.217948] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 839.218307] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e24560b1-a2c2-49b4-bad0-4ff1184dbcc0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.225614] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 839.225614] env[66641]: value = "task-5146136" [ 839.225614] env[66641]: _type = "Task" [ 839.225614] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.239284] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.271563] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 839.293137] env[66641]: INFO nova.compute.manager [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Rebuilding instance [ 839.352017] env[66641]: DEBUG nova.compute.manager [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 839.352985] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e5f69d-d9c6-40d7-8a13-8e8db1333c52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.455725] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146135, 'name': RemoveSnapshot_Task} progress is 98%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.587328] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2eb2ac6-571b-4da9-8b39-1850d8d69601 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.602270] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9de61cdd-8745-450c-8149-7cc7adadf219 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.628806] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52812730-c023-94b4-ae60-3d174662f5cd, 'name': SearchDatastore_Task, 'duration_secs': 0.013217} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.629254] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.629463] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.629615] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.646079] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a651c5c-ad56-4612-a3b7-014192163ab7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.649546] env[66641]: DEBUG nova.compute.manager [req-54a3991c-1df6-4d3a-9fc9-c9d10b4db2de req-e5e9b3cd-7b9d-453d-b5e3-5cee4e1da44f service nova] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Detach interface failed, port_id=8b30e411-0e23-4c9f-bfe8-eb80a3cbd6d1, reason: Instance acde41f9-4256-4e63-98e3-fd092c66c71e could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 839.656083] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52484c5a-ab0e-5395-c934-75652a54cfa4, 'name': SearchDatastore_Task, 'duration_secs': 0.013264} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.658882] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d7571869-794e-453f-85f7-f204d4e1e036 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.663103] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9fd8162-4372-4079-bc57-9731acb3ba2e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.671409] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 839.671409] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231bf17-2025-d951-bd37-f4d2ec7bae21" [ 839.671409] env[66641]: _type = "Task" [ 839.671409] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.704057] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43a29df8-e306-42e3-bd3e-1665f84e3089 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.719024] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df6d3f4-2c38-4749-8c68-d6e2dd8ed537 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.720028] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5231bf17-2025-d951-bd37-f4d2ec7bae21, 'name': SearchDatastore_Task, 'duration_secs': 0.012452} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.720676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 839.720676] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 4e70780a-62ca-4e4a-9366-00dc35750c61/4e70780a-62ca-4e4a-9366-00dc35750c61.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.721380] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 839.721743] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.722186] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-771a098c-3087-4767-8150-b2f5343079b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.735931] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-abbdd81b-ca7c-4faa-b44c-16b2203d92ac {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.739270] env[66641]: DEBUG nova.compute.provider_tree [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.747632] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 839.747632] env[66641]: value = "task-5146137" [ 839.747632] env[66641]: _type = "Task" [ 839.747632] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.751171] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146136, 'name': PowerOffVM_Task, 'duration_secs': 0.474484} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.755641] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 839.755835] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 839.756152] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.756314] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 839.757700] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af1beb03-8946-46b4-a02d-ac2272f31a26 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.759394] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62919c92-051c-449d-9a7f-013faebe5b55 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.769937] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146137, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.770079] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 839.770079] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f60692-fa48-3f85-4afc-cc84340c821e" [ 839.770079] env[66641]: _type = "Task" [ 839.770079] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.780316] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52f60692-fa48-3f85-4afc-cc84340c821e, 'name': SearchDatastore_Task, 'duration_secs': 0.01092} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.781150] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22b2bb37-cd6c-459e-8922-f82e6c523f03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.788426] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 839.788426] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5287a610-8112-2077-501a-755b30cac395" [ 839.788426] env[66641]: _type = "Task" [ 839.788426] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.798226] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5287a610-8112-2077-501a-755b30cac395, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.836372] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 839.836581] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 839.836774] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleting the datastore file [datastore1] dfa8c73b-db57-42a9-a9a4-cf812f5b2949 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 839.837130] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2310c8cc-6ee4-42b4-b07c-cc09b8ce7604 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.845228] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for the task: (returnval){ [ 839.845228] env[66641]: value = "task-5146139" [ 839.845228] env[66641]: _type = "Task" [ 839.845228] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.854257] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146139, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.959497] env[66641]: DEBUG oslo_vmware.api [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146135, 'name': RemoveSnapshot_Task, 'duration_secs': 1.747675} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.959831] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 839.960098] env[66641]: INFO nova.compute.manager [None req-8a7ce599-4ace-4740-80b3-cae0a067c361 tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 16.90 seconds to snapshot the instance on the hypervisor. [ 840.247272] env[66641]: DEBUG nova.scheduler.client.report [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 840.267610] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146137, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.300410] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5287a610-8112-2077-501a-755b30cac395, 'name': SearchDatastore_Task, 'duration_secs': 0.01182} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.301021] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 840.301421] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 29f03150-f0a2-4424-b758-32891730923c/29f03150-f0a2-4424-b758-32891730923c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.302279] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-523d7176-69fd-4085-8af0-d28501e325f0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.310759] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 840.310759] env[66641]: value = "task-5146140" [ 840.310759] env[66641]: _type = "Task" [ 840.310759] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.320968] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146140, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.357236] env[66641]: DEBUG oslo_vmware.api [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Task: {'id': task-5146139, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367866} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.357788] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.358137] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 840.358521] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 840.358886] env[66641]: INFO nova.compute.manager [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Took 1.15 seconds to destroy the instance on the hypervisor. [ 840.359311] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 840.360124] env[66641]: DEBUG nova.compute.manager [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 840.360124] env[66641]: DEBUG nova.network.neutron [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 840.360657] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.361120] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.369200] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.369476] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ae4469a-1123-434e-a64b-db8e6eaa384f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.378056] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 840.378056] env[66641]: value = "task-5146141" [ 840.378056] env[66641]: _type = "Task" [ 840.378056] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.388349] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146141, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.405269] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 840.405651] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 840.728029] env[66641]: DEBUG nova.compute.manager [req-b938bc64-a31b-42f9-9e43-4df726c257dd req-f60d2787-3390-4717-ad32-59a59c968014 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Received event network-vif-deleted-5ebec271-9d57-41f2-ba8c-59edb82c50db {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 840.728248] env[66641]: INFO nova.compute.manager [req-b938bc64-a31b-42f9-9e43-4df726c257dd req-f60d2787-3390-4717-ad32-59a59c968014 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Neutron deleted interface 5ebec271-9d57-41f2-ba8c-59edb82c50db; detaching it from the instance and deleting it from the info cache [ 840.728390] env[66641]: DEBUG nova.network.neutron [req-b938bc64-a31b-42f9-9e43-4df726c257dd req-f60d2787-3390-4717-ad32-59a59c968014 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 840.752469] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.799s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 840.756269] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.485s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 840.756537] env[66641]: DEBUG nova.objects.instance [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lazy-loading 'resources' on Instance uuid e508b1e9-cbb0-408c-98e1-d54ceaec1f1b {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 840.771936] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146137, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526626} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.772348] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 4e70780a-62ca-4e4a-9366-00dc35750c61/4e70780a-62ca-4e4a-9366-00dc35750c61.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.772544] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.773792] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-414ecc04-4f2e-4420-8f91-9290d396fea5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.786183] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 840.786183] env[66641]: value = "task-5146142" [ 840.786183] env[66641]: _type = "Task" [ 840.786183] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.793837] env[66641]: INFO nova.scheduler.client.report [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Deleted allocations for instance acde41f9-4256-4e63-98e3-fd092c66c71e [ 840.803037] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146142, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.826154] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146140, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.892029] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146141, 'name': PowerOffVM_Task, 'duration_secs': 0.243244} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.893129] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 840.893724] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.898023] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328a5b5d-3212-4425-9f35-8791d024bbca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.904789] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 840.906166] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3a9b9d64-e688-4f19-be4c-98462ba06adf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.935030] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 840.935299] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 840.935478] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Deleting the datastore file [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.935774] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9f9264f0-d4f5-4978-b386-0e0f95678707 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.943795] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 840.943795] env[66641]: value = "task-5146144" [ 840.943795] env[66641]: _type = "Task" [ 840.943795] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.953486] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146144, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.197147] env[66641]: DEBUG nova.network.neutron [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 841.234543] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7c138cb5-81c1-4e0b-8b2b-d24b6a31c6dd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.246646] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564bcba6-2c33-474b-8224-3965568c0dd9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.285634] env[66641]: DEBUG nova.compute.manager [req-b938bc64-a31b-42f9-9e43-4df726c257dd req-f60d2787-3390-4717-ad32-59a59c968014 service nova] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Detach interface failed, port_id=5ebec271-9d57-41f2-ba8c-59edb82c50db, reason: Instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 841.299762] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146142, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079573} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.299890] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.300690] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77fb673e-b40d-4bf4-adc3-adcd1de28849 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.308247] env[66641]: DEBUG oslo_concurrency.lockutils [None req-bca32b03-13ce-42c5-b270-c39d49ff30ec tempest-AttachVolumeTestJSON-1398173537 tempest-AttachVolumeTestJSON-1398173537-project-member] Lock "acde41f9-4256-4e63-98e3-fd092c66c71e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.847s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.332303] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] 4e70780a-62ca-4e4a-9366-00dc35750c61/4e70780a-62ca-4e4a-9366-00dc35750c61.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.339920] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b2301fe-c55a-4314-845d-02596dda4242 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.367179] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146140, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518643} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.368668] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 29f03150-f0a2-4424-b758-32891730923c/29f03150-f0a2-4424-b758-32891730923c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.368900] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.369251] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 841.369251] env[66641]: value = "task-5146145" [ 841.369251] env[66641]: _type = "Task" [ 841.369251] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.369447] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bcd26121-7c07-44bf-be07-361011fa8091 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.385619] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 841.385619] env[66641]: value = "task-5146146" [ 841.385619] env[66641]: _type = "Task" [ 841.385619] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.398871] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146146, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.456265] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146144, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103921} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.459922] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.459922] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.459922] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.481626] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "40d7571f-09e2-463a-a449-36c621045819" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.481916] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.482220] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "40d7571f-09e2-463a-a449-36c621045819-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 841.483031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 841.483031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 841.484709] env[66641]: INFO nova.compute.manager [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Terminating instance [ 841.527174] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304790f3-cb8e-404f-9be7-77a8382f640b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.537753] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-832848d3-066b-4273-9200-8395e29a17f7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.570306] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a601c366-ffd4-49db-b626-7e67df881482 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.578700] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4986a2-bba2-43f9-bef0-6ddc0ca57f89 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.593695] env[66641]: DEBUG nova.compute.provider_tree [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.701824] env[66641]: INFO nova.compute.manager [-] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Took 1.34 seconds to deallocate network for instance. [ 841.885107] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146145, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.894824] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146146, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120602} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.895199] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.895988] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0bbd15-3523-4460-8428-1fa050fcc532 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.918970] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 29f03150-f0a2-4424-b758-32891730923c/29f03150-f0a2-4424-b758-32891730923c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.922321] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21f9749c-db7f-496d-920f-b75d84f2e691 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.943963] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 841.943963] env[66641]: value = "task-5146147" [ 841.943963] env[66641]: _type = "Task" [ 841.943963] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.953238] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146147, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.988665] env[66641]: DEBUG nova.compute.manager [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 841.988927] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 841.989807] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4a7e479-e7b4-404a-bb6e-f9abb905f356 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.998405] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 841.999325] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-990ab11a-e071-4814-9bd7-05f5927695d1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.005074] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 842.005074] env[66641]: value = "task-5146148" [ 842.005074] env[66641]: _type = "Task" [ 842.005074] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.015803] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.097406] env[66641]: DEBUG nova.scheduler.client.report [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 842.213610] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 842.383875] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146145, 'name': ReconfigVM_Task, 'duration_secs': 0.702131} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.384317] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Reconfigured VM instance instance-0000002f to attach disk [datastore2] 4e70780a-62ca-4e4a-9366-00dc35750c61/4e70780a-62ca-4e4a-9366-00dc35750c61.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.384884] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e2d022f8-7fbf-4f33-b1ca-e6909a2e549e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.393068] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 842.393068] env[66641]: value = "task-5146149" [ 842.393068] env[66641]: _type = "Task" [ 842.393068] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.409671] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146149, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.455902] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146147, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.498982] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 842.499510] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 842.499803] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 842.500042] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 842.500259] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 842.500518] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 842.500780] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.501212] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 842.501350] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 842.501596] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 842.502029] env[66641]: DEBUG nova.virt.hardware [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 842.503237] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3892f9a-5c7c-4d75-8eec-c121c04420d3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.520726] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.522438] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90f3a00c-61d5-4720-990d-142fa3e3690a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.540871] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.547462] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 842.547730] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.547961] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-07fd379e-f155-44f1-a193-ffce7fb88e16 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.567736] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.567736] env[66641]: value = "task-5146150" [ 842.567736] env[66641]: _type = "Task" [ 842.567736] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.578545] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146150, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.603676] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.847s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 842.606413] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.393s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 842.606670] env[66641]: DEBUG nova.objects.instance [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lazy-loading 'resources' on Instance uuid dfa8c73b-db57-42a9-a9a4-cf812f5b2949 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 842.636290] env[66641]: INFO nova.scheduler.client.report [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleted allocations for instance e508b1e9-cbb0-408c-98e1-d54ceaec1f1b [ 842.907046] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146149, 'name': Rename_Task, 'duration_secs': 0.153361} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.907597] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.908055] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4edbc160-53e7-4ddd-820a-2245a93bbf76 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.921031] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 842.921031] env[66641]: value = "task-5146151" [ 842.921031] env[66641]: _type = "Task" [ 842.921031] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.933053] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146151, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.957330] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146147, 'name': ReconfigVM_Task, 'duration_secs': 0.885654} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.957590] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 29f03150-f0a2-4424-b758-32891730923c/29f03150-f0a2-4424-b758-32891730923c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.958262] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24ebcc35-b702-4afd-8fac-878485454986 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.965288] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 842.965288] env[66641]: value = "task-5146152" [ 842.965288] env[66641]: _type = "Task" [ 842.965288] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.978668] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146152, 'name': Rename_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.017919] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146148, 'name': PowerOffVM_Task, 'duration_secs': 0.709884} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.018889] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 843.019203] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 843.019418] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-caf48e71-5ccf-41f9-a0e7-4ee0d14c46e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.084421] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146150, 'name': CreateVM_Task, 'duration_secs': 0.299129} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.084421] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.088687] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.088687] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.088687] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 843.088687] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccf8ffdb-65f9-43da-a9f8-04c3015a82cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.093441] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 843.093441] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52661091-e4bb-7bdf-c32e-8548b3f473a3" [ 843.093441] env[66641]: _type = "Task" [ 843.093441] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.101020] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 843.101020] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 843.101020] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Deleting the datastore file [datastore1] 40d7571f-09e2-463a-a449-36c621045819 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.101576] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14944f61-4541-4237-a2ba-6e2a47ebc465 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.108537] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52661091-e4bb-7bdf-c32e-8548b3f473a3, 'name': SearchDatastore_Task, 'duration_secs': 0.012152} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.109024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.109663] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.109873] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.111066] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 843.111066] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.112325] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-45423a9f-0af1-43dd-899f-e2d47f8528fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.117572] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for the task: (returnval){ [ 843.117572] env[66641]: value = "task-5146154" [ 843.117572] env[66641]: _type = "Task" [ 843.117572] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.129149] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146154, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.136124] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 843.136521] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 843.137348] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80b4f31e-e518-4091-8902-92349f6df7c7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.145769] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 843.145769] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5215aa10-ccf3-11e6-0c26-22891a24e7f9" [ 843.145769] env[66641]: _type = "Task" [ 843.145769] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.146799] env[66641]: DEBUG oslo_concurrency.lockutils [None req-736dfd54-03ce-4531-ac2e-35d6c286df39 tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "e508b1e9-cbb0-408c-98e1-d54ceaec1f1b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.561s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 843.163073] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5215aa10-ccf3-11e6-0c26-22891a24e7f9, 'name': SearchDatastore_Task, 'duration_secs': 0.010141} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.166684] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e14b438-fd86-41ca-a83b-97dc2a7e6da8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.173623] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 843.173623] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ea0a9e-ceed-98ae-2cae-7128b899d21f" [ 843.173623] env[66641]: _type = "Task" [ 843.173623] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.186618] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ea0a9e-ceed-98ae-2cae-7128b899d21f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.286814] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-760a6e18-dde8-4506-af5c-86dc7252a8ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.294717] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86314b54-6cef-49b4-9609-f5cbe74db21a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.330163] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-200cfc57-c2de-4e32-82cc-200473215d17 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.340031] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd9eea7d-3837-4204-ad02-97e73c12e069 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.354709] env[66641]: DEBUG nova.compute.provider_tree [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 843.433046] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146151, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.475383] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146152, 'name': Rename_Task, 'duration_secs': 0.174847} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.475674] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 843.476008] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-78c325d9-ff1c-42cd-a3f4-6c5e0bfdcfdb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.483422] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 843.483422] env[66641]: value = "task-5146155" [ 843.483422] env[66641]: _type = "Task" [ 843.483422] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.493717] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146155, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.628795] env[66641]: DEBUG oslo_vmware.api [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Task: {'id': task-5146154, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175186} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.629131] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 843.629299] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 843.629472] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 843.629748] env[66641]: INFO nova.compute.manager [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 1.64 seconds to destroy the instance on the hypervisor. [ 843.629955] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 843.630286] env[66641]: DEBUG nova.compute.manager [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 843.630336] env[66641]: DEBUG nova.network.neutron [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 843.630837] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.631124] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.688766] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52ea0a9e-ceed-98ae-2cae-7128b899d21f, 'name': SearchDatastore_Task, 'duration_secs': 0.010758} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.692471] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 843.692744] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 843.700550] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 843.700832] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 843.701377] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-54928095-96c2-4417-9dd6-18910713851a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.711987] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 843.711987] env[66641]: value = "task-5146156" [ 843.711987] env[66641]: _type = "Task" [ 843.711987] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.723655] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146156, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.858507] env[66641]: DEBUG nova.scheduler.client.report [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 843.935605] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146151, 'name': PowerOnVM_Task} progress is 66%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.011578] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146155, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.061244] env[66641]: DEBUG nova.compute.manager [req-1cee0e90-c75b-4f51-91f5-a1c1a990d274 req-9acbdc8d-cc96-4ea8-b92d-e591b167fe62 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Received event network-vif-deleted-5f780526-bca9-4384-a9fc-e10c3d80fb62 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 844.061398] env[66641]: INFO nova.compute.manager [req-1cee0e90-c75b-4f51-91f5-a1c1a990d274 req-9acbdc8d-cc96-4ea8-b92d-e591b167fe62 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Neutron deleted interface 5f780526-bca9-4384-a9fc-e10c3d80fb62; detaching it from the instance and deleting it from the info cache [ 844.061604] env[66641]: DEBUG nova.network.neutron [req-1cee0e90-c75b-4f51-91f5-a1c1a990d274 req-9acbdc8d-cc96-4ea8-b92d-e591b167fe62 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.222703] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146156, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.366120] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.760s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.394719] env[66641]: INFO nova.scheduler.client.report [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Deleted allocations for instance dfa8c73b-db57-42a9-a9a4-cf812f5b2949 [ 844.433397] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146151, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.495922] env[66641]: DEBUG oslo_vmware.api [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146155, 'name': PowerOnVM_Task, 'duration_secs': 0.661584} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.496288] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.496503] env[66641]: INFO nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Took 9.73 seconds to spawn the instance on the hypervisor. [ 844.496685] env[66641]: DEBUG nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 844.497601] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fd66bac-7633-44f0-be7c-4d6fe9a5912f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.500728] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.500920] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.501160] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 844.501398] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 844.501507] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.508774] env[66641]: INFO nova.compute.manager [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Terminating instance [ 844.533856] env[66641]: DEBUG nova.network.neutron [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 844.566092] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b235508-3c47-42bc-94ce-20676351327c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.579840] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b084371-f653-41ac-89d5-edcac9a49145 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.614924] env[66641]: DEBUG nova.compute.manager [req-1cee0e90-c75b-4f51-91f5-a1c1a990d274 req-9acbdc8d-cc96-4ea8-b92d-e591b167fe62 service nova] [instance: 40d7571f-09e2-463a-a449-36c621045819] Detach interface failed, port_id=5f780526-bca9-4384-a9fc-e10c3d80fb62, reason: Instance 40d7571f-09e2-463a-a449-36c621045819 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 844.724035] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146156, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620261} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.724433] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 844.724433] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 844.724740] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-10ac06f1-d6dd-4c3f-aca9-be01045dd70f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.733373] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 844.733373] env[66641]: value = "task-5146157" [ 844.733373] env[66641]: _type = "Task" [ 844.733373] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.745790] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146157, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.902808] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ac0d626a-5bda-49e9-ab6f-5b8e293d0c62 tempest-MigrationsAdminTest-712875926 tempest-MigrationsAdminTest-712875926-project-member] Lock "dfa8c73b-db57-42a9-a9a4-cf812f5b2949" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.206s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 844.934035] env[66641]: DEBUG oslo_vmware.api [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146151, 'name': PowerOnVM_Task, 'duration_secs': 1.726264} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.934776] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 844.934776] env[66641]: INFO nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Took 12.22 seconds to spawn the instance on the hypervisor. [ 844.934776] env[66641]: DEBUG nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 844.935639] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d0af56f-303c-4a11-9bff-190ddc55ecef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.021071] env[66641]: DEBUG nova.compute.manager [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 845.021071] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.022532] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012480f1-8627-42c6-a46a-649bdf8cb666 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.027475] env[66641]: INFO nova.compute.manager [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Took 23.54 seconds to build instance. [ 845.034864] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.035320] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3369a5bd-115c-41c5-af90-7dfd27f53ba2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.041086] env[66641]: INFO nova.compute.manager [-] [instance: 40d7571f-09e2-463a-a449-36c621045819] Took 1.41 seconds to deallocate network for instance. [ 845.053618] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 845.053618] env[66641]: value = "task-5146158" [ 845.053618] env[66641]: _type = "Task" [ 845.053618] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.063023] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.247647] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146157, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.212277} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.248092] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 845.249198] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d1b4e60-57fc-4d8c-b95c-6249a5054892 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.276479] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 845.276841] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f8f641e5-8544-4042-8697-e7b4dcdd1565 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.299682] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 845.299682] env[66641]: value = "task-5146159" [ 845.299682] env[66641]: _type = "Task" [ 845.299682] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.312025] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146159, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.458187] env[66641]: INFO nova.compute.manager [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Took 24.98 seconds to build instance. [ 845.529336] env[66641]: DEBUG oslo_concurrency.lockutils [None req-fee51b33-abcd-4060-bbe3-2ae02f045b98 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.056s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.545030] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "29f03150-f0a2-4424-b758-32891730923c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.546115] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.546386] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "29f03150-f0a2-4424-b758-32891730923c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.546572] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.546741] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 845.553748] env[66641]: INFO nova.compute.manager [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Terminating instance [ 845.556401] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 845.556632] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 845.556960] env[66641]: DEBUG nova.objects.instance [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lazy-loading 'resources' on Instance uuid 40d7571f-09e2-463a-a449-36c621045819 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.574395] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146158, 'name': PowerOffVM_Task, 'duration_secs': 0.405785} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.575607] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 845.575819] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 845.577174] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b9a7b09-78a0-4737-8e15-66873e8c644a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.655374] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 845.659022] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 845.659264] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleting the datastore file [datastore2] a7575399-f72d-4d2c-b57b-f2f2f591c8a7 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 845.659584] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-473669d1-4f2e-4060-b6e7-62f4be85272d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.669365] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for the task: (returnval){ [ 845.669365] env[66641]: value = "task-5146161" [ 845.669365] env[66641]: _type = "Task" [ 845.669365] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.684201] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.810522] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146159, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.960774] env[66641]: DEBUG oslo_concurrency.lockutils [None req-97873af0-7681-442b-8ab6-81994011cf13 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.518s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 846.069622] env[66641]: DEBUG nova.compute.manager [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 846.069841] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 846.071513] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345ae47c-ee43-4db1-a719-23f36e86a317 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.080760] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 846.081056] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6c8dc4e6-0895-45a4-ae57-08681e193693 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.090496] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 846.090496] env[66641]: value = "task-5146162" [ 846.090496] env[66641]: _type = "Task" [ 846.090496] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.101644] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146162, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.182624] env[66641]: DEBUG oslo_vmware.api [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Task: {'id': task-5146161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.21136} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.182851] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.183047] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.183235] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.183434] env[66641]: INFO nova.compute.manager [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 1.16 seconds to destroy the instance on the hypervisor. [ 846.183805] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 846.184069] env[66641]: DEBUG nova.compute.manager [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 846.184230] env[66641]: DEBUG nova.network.neutron [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 846.184772] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.185053] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.244585] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 846.244942] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 846.255250] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f6c57b-a43e-44a4-b8df-8d4b54eea9a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.267310] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c305edc-1bec-4578-ae92-7aac1784b4d0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.302877] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "4e70780a-62ca-4e4a-9366-00dc35750c61" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 846.302877] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 846.302877] env[66641]: INFO nova.compute.manager [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Rebooting instance [ 846.310904] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cff6d4-4af7-4af0-8cdf-70a80e28363f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.323518] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146159, 'name': ReconfigVM_Task, 'duration_secs': 0.628085} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.325772] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Reconfigured VM instance instance-0000002e to attach disk [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527/a4f27205-f8ce-49f9-a3de-2e53bcbb6527.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 846.326543] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc017490-c22b-45bb-9eee-f8577862ec7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.329577] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4887e8a2-a709-48a3-840f-45fe01b3cb8a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.347161] env[66641]: DEBUG nova.compute.provider_tree [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.349894] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 846.349894] env[66641]: value = "task-5146163" [ 846.349894] env[66641]: _type = "Task" [ 846.349894] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.359098] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146163, 'name': Rename_Task} progress is 10%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.413448] env[66641]: DEBUG nova.compute.manager [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 846.414538] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2984c663-870b-4c0c-8450-648724467056 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.518422] env[66641]: DEBUG nova.compute.manager [req-b4398f5c-cee4-43cd-ab57-d38cc18bf9d0 req-d8eda678-7b26-4177-8635-b16190bab24b service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Received event network-vif-deleted-ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 846.518603] env[66641]: INFO nova.compute.manager [req-b4398f5c-cee4-43cd-ab57-d38cc18bf9d0 req-d8eda678-7b26-4177-8635-b16190bab24b service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Neutron deleted interface ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6; detaching it from the instance and deleting it from the info cache [ 846.518769] env[66641]: DEBUG nova.network.neutron [req-b4398f5c-cee4-43cd-ab57-d38cc18bf9d0 req-d8eda678-7b26-4177-8635-b16190bab24b service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 846.603681] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146162, 'name': PowerOffVM_Task, 'duration_secs': 0.298036} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.604105] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.604432] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.604696] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-95fc3ecd-2e60-483e-b7b7-8eabfe866f79 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.690277] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.690277] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.691357] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Deleting the datastore file [datastore2] 29f03150-f0a2-4424-b758-32891730923c {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.691357] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba36d389-c42e-4732-afaf-3f4c709a99f3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.700842] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for the task: (returnval){ [ 846.700842] env[66641]: value = "task-5146165" [ 846.700842] env[66641]: _type = "Task" [ 846.700842] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.711747] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146165, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.839909] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.840153] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquired lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 846.840368] env[66641]: DEBUG nova.network.neutron [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 846.851975] env[66641]: DEBUG nova.scheduler.client.report [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 846.870121] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146163, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.925565] env[66641]: INFO nova.compute.manager [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] instance snapshotting [ 846.926230] env[66641]: DEBUG nova.objects.instance [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lazy-loading 'flavor' on Instance uuid c95b481a-7956-410d-971c-7d94911230bb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 846.974037] env[66641]: DEBUG nova.network.neutron [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 847.023027] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dee97bfc-d2ac-4e78-a858-e01d6d922d2a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.035522] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22140394-8340-433d-a0c1-7687eb54e947 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.068707] env[66641]: DEBUG nova.compute.manager [req-b4398f5c-cee4-43cd-ab57-d38cc18bf9d0 req-d8eda678-7b26-4177-8635-b16190bab24b service nova] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Detach interface failed, port_id=ca1fff4b-fcc8-43e8-ba53-a6cd3329cba6, reason: Instance a7575399-f72d-4d2c-b57b-f2f2f591c8a7 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 847.214139] env[66641]: DEBUG oslo_vmware.api [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Task: {'id': task-5146165, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.293396} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.214139] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.214358] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.214493] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.214664] env[66641]: INFO nova.compute.manager [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] [instance: 29f03150-f0a2-4424-b758-32891730923c] Took 1.14 seconds to destroy the instance on the hypervisor. [ 847.214912] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 847.215131] env[66641]: DEBUG nova.compute.manager [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 847.215265] env[66641]: DEBUG nova.network.neutron [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 847.215755] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.216013] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.268256] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.268550] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.344146] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.344574] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.363517] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.369868] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146163, 'name': Rename_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.407122] env[66641]: INFO nova.scheduler.client.report [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Deleted allocations for instance 40d7571f-09e2-463a-a449-36c621045819 [ 847.436539] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a07b2f5-e418-4e8e-bb8a-eb7cbc43235a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.466233] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab6c07f-7809-4c77-9467-b97a2d8a344c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.478156] env[66641]: INFO nova.compute.manager [-] [instance: a7575399-f72d-4d2c-b57b-f2f2f591c8a7] Took 1.29 seconds to deallocate network for instance. [ 847.637025] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.637825] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.718546] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 847.719092] env[66641]: WARNING openstack [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 847.869880] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146163, 'name': Rename_Task, 'duration_secs': 1.272872} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.870232] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 847.870551] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ef2930c1-f338-459e-b7df-b8f982b41a21 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.878500] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 847.878500] env[66641]: value = "task-5146166" [ 847.878500] env[66641]: _type = "Task" [ 847.878500] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.888232] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146166, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.916699] env[66641]: DEBUG oslo_concurrency.lockutils [None req-11c91d6f-48ed-44b5-93c3-36594b8aa04a tempest-ImagesOneServerTestJSON-2141454217 tempest-ImagesOneServerTestJSON-2141454217-project-member] Lock "40d7571f-09e2-463a-a449-36c621045819" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.435s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 847.982400] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 847.982750] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bc912b54-96a8-47f9-8769-8b7ae01685b5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.987332] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 847.988496] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 847.988496] env[66641]: DEBUG nova.objects.instance [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lazy-loading 'resources' on Instance uuid a7575399-f72d-4d2c-b57b-f2f2f591c8a7 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.994813] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 847.994813] env[66641]: value = "task-5146167" [ 847.994813] env[66641]: _type = "Task" [ 847.994813] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.997556] env[66641]: DEBUG nova.network.neutron [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Updating instance_info_cache with network_info: [{"id": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "address": "fa:16:3e:fc:7c:97", "network": {"id": "278f3697-035d-48dc-9175-7202147bad69", "bridge": "br-int", "label": "tempest-InstanceActionsTestJSON-1403293202-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "26e21ce2b8bc4f209cc4ac5453d151e7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d69a4b11-8d65-435f-94a5-28f74a39a718", "external-id": "cl2-zone-59", "segmentation_id": 59, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa48fa631-89", "ovs_interfaceid": "a48fa631-89c4-45c7-bb90-dce5ffb9e9c6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.010356] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146167, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.059126] env[66641]: DEBUG nova.network.neutron [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 848.393181] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146166, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.504166] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Releasing lock "refresh_cache-4e70780a-62ca-4e4a-9366-00dc35750c61" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 848.515621] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146167, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.552668] env[66641]: DEBUG nova.compute.manager [req-79a46e3f-d482-4491-84e5-cb301fad77ed req-7ae3dfaf-bb79-453e-875e-ddec5307f8ad service nova] [instance: 29f03150-f0a2-4424-b758-32891730923c] Received event network-vif-deleted-4100a6ee-e7e0-4d1f-afe1-05af072a5ce8 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 848.565921] env[66641]: INFO nova.compute.manager [-] [instance: 29f03150-f0a2-4424-b758-32891730923c] Took 1.35 seconds to deallocate network for instance. [ 848.656291] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf703fda-9f78-45ab-a03b-bd5828887025 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.673861] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b14d6c5b-cde4-461f-b3c5-a2adc6c4c4fd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.712493] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09aacf25-984d-409a-87ea-b691568f6695 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.726995] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1e285b-1ab2-4946-bd2b-2f074c88d95d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.742186] env[66641]: DEBUG nova.compute.provider_tree [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.892809] env[66641]: DEBUG oslo_vmware.api [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146166, 'name': PowerOnVM_Task, 'duration_secs': 0.59928} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.893312] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 848.893477] env[66641]: DEBUG nova.compute.manager [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 848.894454] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66e7046e-02fe-4f3b-a8d8-b19391d0d760 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.015910] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146167, 'name': CreateSnapshot_Task, 'duration_secs': 0.547209} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.016230] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 849.017482] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-013376b4-0100-45cd-a63a-a2a1e698fea6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.019957] env[66641]: DEBUG nova.compute.manager [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 849.020715] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c59173-f077-4b17-9158-40d1c55e867b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.074445] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.185680] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 849.185969] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Cleaning up deleted instances {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11932}} [ 849.247126] env[66641]: DEBUG nova.scheduler.client.report [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 849.414174] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 849.543758] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 849.546898] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f25c899a-d9e3-4305-ad89-fbfd759a1107 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.560772] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 849.560772] env[66641]: value = "task-5146171" [ 849.560772] env[66641]: _type = "Task" [ 849.560772] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.570514] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146171, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.705346] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] There are 40 instances to clean {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11941}} [ 849.705740] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: acde41f9-4256-4e63-98e3-fd092c66c71e] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 849.753113] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.765s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 849.762722] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.688s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 849.762811] env[66641]: DEBUG nova.objects.instance [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lazy-loading 'resources' on Instance uuid 29f03150-f0a2-4424-b758-32891730923c {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.800707] env[66641]: INFO nova.scheduler.client.report [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Deleted allocations for instance a7575399-f72d-4d2c-b57b-f2f2f591c8a7 [ 850.053138] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d77ff699-bafa-4319-b210-ea547b515f5f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.061646] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Doing hard reboot of VM {{(pid=66641) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 850.065764] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-082f727c-e39b-4d13-8e85-27a6a3ef7ba4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.075175] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146171, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.076539] env[66641]: DEBUG oslo_vmware.api [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 850.076539] env[66641]: value = "task-5146172" [ 850.076539] env[66641]: _type = "Task" [ 850.076539] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.121448] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.121792] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.122212] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 850.122447] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 850.122616] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.125305] env[66641]: INFO nova.compute.manager [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Terminating instance [ 850.211453] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 40d7571f-09e2-463a-a449-36c621045819] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 850.311834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-31d7d6d2-e87d-4744-96da-75a9566a8bff tempest-ListImageFiltersTestJSON-497591182 tempest-ListImageFiltersTestJSON-497591182-project-member] Lock "a7575399-f72d-4d2c-b57b-f2f2f591c8a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.811s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 850.389101] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2007401d-2217-493a-8a2c-8ce64fec84af {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.398731] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59648fb4-0960-4074-a687-a459422cfd2f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.437142] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c598e2d3-bb1a-4d9d-bd4a-a690cabbb980 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.446926] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a852cb7-5370-4c20-b6c7-cbeb024933f1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.462633] env[66641]: DEBUG nova.compute.provider_tree [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.572278] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146171, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.589014] env[66641]: DEBUG oslo_vmware.api [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146172, 'name': ResetVM_Task, 'duration_secs': 0.129617} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.589233] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Did hard reboot of VM {{(pid=66641) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 850.589398] env[66641]: DEBUG nova.compute.manager [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 850.590248] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d06ae6-9eeb-4031-810d-90aedee0a784 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.631242] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "refresh_cache-a4f27205-f8ce-49f9-a3de-2e53bcbb6527" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.631242] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquired lock "refresh_cache-a4f27205-f8ce-49f9-a3de-2e53bcbb6527" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 850.631242] env[66641]: DEBUG nova.network.neutron [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 850.715322] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 207483d3-803c-495b-9b93-6f986f3ca56e] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 850.966741] env[66641]: DEBUG nova.scheduler.client.report [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 851.073826] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146171, 'name': CloneVM_Task} progress is 95%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.106879] env[66641]: DEBUG oslo_concurrency.lockutils [None req-d09ee8da-bd49-47ca-9c5d-fc6ed4d7c0ff tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.804s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.135292] env[66641]: WARNING openstack [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 851.135741] env[66641]: WARNING openstack [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 851.140848] env[66641]: DEBUG nova.network.neutron [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 851.215884] env[66641]: DEBUG nova.network.neutron [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 851.217945] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 34a98372-2ab7-4b21-8a0e-2fc3b91ef4db] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 851.473477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.711s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 851.479202] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.065s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 851.479827] env[66641]: DEBUG nova.objects.instance [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Trying to apply a migration context that does not seem to be set for this instance {{(pid=66641) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 851.517627] env[66641]: INFO nova.scheduler.client.report [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Deleted allocations for instance 29f03150-f0a2-4424-b758-32891730923c [ 851.579643] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146171, 'name': CloneVM_Task, 'duration_secs': 1.713959} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.580170] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created linked-clone VM from snapshot [ 851.581323] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339fc89d-fd01-4ea0-ba78-18cc4e804f4b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.593616] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Uploading image ccac96b4-e93b-428d-b8e5-d64b8daec92a {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 851.629165] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 851.629165] env[66641]: value = "vm-1000721" [ 851.629165] env[66641]: _type = "VirtualMachine" [ 851.629165] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 851.629165] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9db5b9bf-5f89-4352-9332-bcd55aed8c3b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.641416] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease: (returnval){ [ 851.641416] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbff8d-960c-2dea-e8b1-dc66ac3bb80b" [ 851.641416] env[66641]: _type = "HttpNfcLease" [ 851.641416] env[66641]: } obtained for exporting VM: (result){ [ 851.641416] env[66641]: value = "vm-1000721" [ 851.641416] env[66641]: _type = "VirtualMachine" [ 851.641416] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 851.641877] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the lease: (returnval){ [ 851.641877] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbff8d-960c-2dea-e8b1-dc66ac3bb80b" [ 851.641877] env[66641]: _type = "HttpNfcLease" [ 851.641877] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 851.655098] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 851.655098] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbff8d-960c-2dea-e8b1-dc66ac3bb80b" [ 851.655098] env[66641]: _type = "HttpNfcLease" [ 851.655098] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 851.655310] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 851.655310] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52cbff8d-960c-2dea-e8b1-dc66ac3bb80b" [ 851.655310] env[66641]: _type = "HttpNfcLease" [ 851.655310] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 851.656126] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a87025-d97b-478d-8ac1-56714696840d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.665465] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 851.665657] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 851.730615] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Releasing lock "refresh_cache-a4f27205-f8ce-49f9-a3de-2e53bcbb6527" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 851.731104] env[66641]: DEBUG nova.compute.manager [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 851.731350] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 851.731856] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 48f99287-b737-45fa-ad59-9e1425afa3d5] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 851.735724] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8368a9fa-7aa7-499f-acac-0629507edcbc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.745546] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 851.745838] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ec89b368-90f9-49bd-a9a0-d87f24c2ffde {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.753126] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 851.753126] env[66641]: value = "task-5146174" [ 851.753126] env[66641]: _type = "Task" [ 851.753126] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.764677] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146174, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.775966] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-853bc425-1bd3-483f-8afd-d36bc8419223 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.029414] env[66641]: DEBUG oslo_concurrency.lockutils [None req-83279de6-0751-402f-b5d9-6a894b761b31 tempest-ServerActionsTestOtherA-1875962759 tempest-ServerActionsTestOtherA-1875962759-project-member] Lock "29f03150-f0a2-4424-b758-32891730923c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.483s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.132957] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "4e70780a-62ca-4e4a-9366-00dc35750c61" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 852.133280] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.133496] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 852.133654] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 852.133834] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.136818] env[66641]: INFO nova.compute.manager [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Terminating instance [ 852.240909] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 081457ae-e152-410c-bca7-4d43b95eee10] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 852.270836] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146174, 'name': PowerOffVM_Task, 'duration_secs': 0.152034} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.271180] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.271358] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.271869] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1e41eff-6118-4f21-875e-762381b4cd13 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.312454] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.312738] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.312951] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Deleting the datastore file [datastore2] a4f27205-f8ce-49f9-a3de-2e53bcbb6527 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.313269] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2a173f3-2a34-4481-9d9f-a13cb95f84d9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.320972] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for the task: (returnval){ [ 852.320972] env[66641]: value = "task-5146176" [ 852.320972] env[66641]: _type = "Task" [ 852.320972] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.331014] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146176, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.493677] env[66641]: DEBUG oslo_concurrency.lockutils [None req-afa89d11-128d-447d-a5e2-03c99bf6e41f tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.015s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 852.642194] env[66641]: DEBUG nova.compute.manager [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 852.642354] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.643333] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da2b0ce4-3315-4416-8fb8-1fa013404957 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.651886] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.652213] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4e55547d-4924-4d82-899e-2f3303f3abb3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.660501] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 852.660501] env[66641]: value = "task-5146178" [ 852.660501] env[66641]: _type = "Task" [ 852.660501] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.669728] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146178, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.745542] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 61042df0-a727-4aa8-b2ea-bdc40899d0fc] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 852.833485] env[66641]: DEBUG oslo_vmware.api [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Task: {'id': task-5146176, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.399975} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.833824] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 852.834145] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 852.834469] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 852.834761] env[66641]: INFO nova.compute.manager [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Took 1.10 seconds to destroy the instance on the hypervisor. [ 852.835069] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 852.835290] env[66641]: DEBUG nova.compute.manager [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 852.835467] env[66641]: DEBUG nova.network.neutron [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 852.836243] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 852.836680] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 852.935902] env[66641]: DEBUG nova.network.neutron [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 852.936658] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 852.936992] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.180132] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146178, 'name': PowerOffVM_Task, 'duration_secs': 0.255982} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.180738] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 853.180738] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 853.181116] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-21c16e93-ec03-4274-ae2d-01c1aa87c5a9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.251086] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: dad5ed23-71a1-4b55-856f-2484f8e62708] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 853.280806] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 853.280806] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 853.280806] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Deleting the datastore file [datastore2] 4e70780a-62ca-4e4a-9366-00dc35750c61 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 853.281183] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-09d83da8-b94c-49d9-8f10-322ea91b4e2e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.290922] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for the task: (returnval){ [ 853.290922] env[66641]: value = "task-5146180" [ 853.290922] env[66641]: _type = "Task" [ 853.290922] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.308206] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146180, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.445256] env[66641]: DEBUG nova.network.neutron [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 853.756896] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 05882781-78be-4568-95f4-2fccc4cf4dfe] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 853.812766] env[66641]: DEBUG oslo_vmware.api [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Task: {'id': task-5146180, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.1944} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.813394] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.813879] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.814330] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.814703] env[66641]: INFO nova.compute.manager [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Took 1.17 seconds to destroy the instance on the hypervisor. [ 853.815076] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 853.815484] env[66641]: DEBUG nova.compute.manager [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 853.815680] env[66641]: DEBUG nova.network.neutron [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 853.816294] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 853.816675] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 853.948620] env[66641]: INFO nova.compute.manager [-] [instance: a4f27205-f8ce-49f9-a3de-2e53bcbb6527] Took 1.11 seconds to deallocate network for instance. [ 853.999699] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 854.000170] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 854.262647] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 25c7bd59-ec24-4d30-840b-3c4549dbf669] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 854.339634] env[66641]: DEBUG nova.compute.manager [req-8c1af20a-0b9c-487a-961f-315f720b0712 req-dbb5cffa-dd71-4c6a-906c-509756a6b44e service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Received event network-vif-deleted-a48fa631-89c4-45c7-bb90-dce5ffb9e9c6 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 854.339634] env[66641]: INFO nova.compute.manager [req-8c1af20a-0b9c-487a-961f-315f720b0712 req-dbb5cffa-dd71-4c6a-906c-509756a6b44e service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Neutron deleted interface a48fa631-89c4-45c7-bb90-dce5ffb9e9c6; detaching it from the instance and deleting it from the info cache [ 854.339823] env[66641]: DEBUG nova.network.neutron [req-8c1af20a-0b9c-487a-961f-315f720b0712 req-dbb5cffa-dd71-4c6a-906c-509756a6b44e service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.458490] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 854.459099] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 854.459396] env[66641]: DEBUG nova.objects.instance [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lazy-loading 'resources' on Instance uuid a4f27205-f8ce-49f9-a3de-2e53bcbb6527 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.762797] env[66641]: DEBUG nova.network.neutron [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 854.766148] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: e508b1e9-cbb0-408c-98e1-d54ceaec1f1b] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 854.842992] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0ba39660-0303-421f-b77a-2c3e1981b60a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.854794] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8ba37fd-a9c9-4f96-81ad-4d61be528eb0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.896418] env[66641]: DEBUG nova.compute.manager [req-8c1af20a-0b9c-487a-961f-315f720b0712 req-dbb5cffa-dd71-4c6a-906c-509756a6b44e service nova] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Detach interface failed, port_id=a48fa631-89c4-45c7-bb90-dce5ffb9e9c6, reason: Instance 4e70780a-62ca-4e4a-9366-00dc35750c61 could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 855.097501] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ec5445-e61a-4a97-86e9-51640ac3898d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.109096] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-291455c4-44fb-4fd5-b07b-78af4470ba3c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.143788] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6365b93-b157-4854-bf28-7dd528b55acc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.155762] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c9a26f0-2d44-43b2-871e-5d720bde3342 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.169994] env[66641]: DEBUG nova.compute.provider_tree [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.268022] env[66641]: INFO nova.compute.manager [-] [instance: 4e70780a-62ca-4e4a-9366-00dc35750c61] Took 1.45 seconds to deallocate network for instance. [ 855.269332] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 0f6055b9-f5b4-48ba-9589-0af212808be7] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 855.674259] env[66641]: DEBUG nova.scheduler.client.report [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 855.774847] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 58fefaa4-0b17-408f-9329-78f8b5cf3fa7] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 855.778754] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 856.179965] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.182759] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.404s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 856.183074] env[66641]: DEBUG nova.objects.instance [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lazy-loading 'resources' on Instance uuid 4e70780a-62ca-4e4a-9366-00dc35750c61 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 856.218641] env[66641]: INFO nova.scheduler.client.report [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Deleted allocations for instance a4f27205-f8ce-49f9-a3de-2e53bcbb6527 [ 856.283334] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 047a5c42-3930-4e6a-b3a5-5dbf55d44a4f] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 856.737795] env[66641]: DEBUG oslo_concurrency.lockutils [None req-91b57f2b-4be3-4743-80e9-fc9fef0e0e2e tempest-ServersListShow296Test-153396115 tempest-ServersListShow296Test-153396115-project-member] Lock "a4f27205-f8ce-49f9-a3de-2e53bcbb6527" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.616s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 856.789206] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: b9d032da-031e-42e0-86e2-95254c1ceac1] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 856.847726] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ddfde5a-4181-48f0-835f-a49432634dc5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.857682] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-718a225e-fe07-4ebd-b4a5-7a08e0b6566d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.901667] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-852e72ee-5b75-43ee-890b-c2eaf5dcc568 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.911853] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0d0cfc-ecc5-4d17-b2c3-af8f23c0720b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.929547] env[66641]: DEBUG nova.compute.provider_tree [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.270812] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 857.270812] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 857.270812] env[66641]: INFO nova.compute.manager [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Shelving [ 857.293998] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 530f1e0e-5911-434e-bf20-edfd9778d7ad] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 857.434479] env[66641]: DEBUG nova.scheduler.client.report [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 857.802185] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: d03f39a4-532b-439a-9055-19fc1e769fff] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 857.945046] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 857.977333] env[66641]: INFO nova.scheduler.client.report [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Deleted allocations for instance 4e70780a-62ca-4e4a-9366-00dc35750c61 [ 858.285054] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 858.285054] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c33c97f9-845f-4f40-b2e1-306fae2ce0b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.292781] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 858.292781] env[66641]: value = "task-5146183" [ 858.292781] env[66641]: _type = "Task" [ 858.292781] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.302579] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146183, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.306603] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 8235441c-c046-49da-a5a5-92bd85e17982] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 858.488179] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b72eb43-b163-48be-a9f6-a97ba70f8084 tempest-InstanceActionsTestJSON-1328448129 tempest-InstanceActionsTestJSON-1328448129-project-member] Lock "4e70780a-62ca-4e4a-9366-00dc35750c61" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.354s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 858.806132] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146183, 'name': PowerOffVM_Task, 'duration_secs': 0.254384} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.806132] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 858.806132] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d66cd1a-b24d-4492-8d67-0de6e7ef0e4b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.825429] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 3619be34-19ed-4d3d-b3ee-573126dcefbb] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 858.828416] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb01742-a74d-42b5-ab8f-388b4068dcb4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.983264] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 858.983682] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 859.332822] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: eeb2556a-4e0c-43d7-83bd-942be5d5fdd2] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 859.342033] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 859.342326] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-a2ff5592-d0cf-4b6f-af7c-ab91a5a115c0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.352636] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 859.352636] env[66641]: value = "task-5146185" [ 859.352636] env[66641]: _type = "Task" [ 859.352636] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.364263] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146185, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.487936] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 859.836597] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: a2d567c9-09af-44e1-aad2-c8f394a09714] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 859.865460] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146185, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.097138] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 860.097477] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 860.099983] env[66641]: INFO nova.compute.claims [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.341899] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: ac5bb5a2-e630-42f0-be0c-24f52be81367] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 860.366517] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146185, 'name': CreateSnapshot_Task, 'duration_secs': 0.883761} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.366817] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 860.367648] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d8c81e-b2bb-479a-8e4f-c4cb35d5ba65 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.845719] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: bc13bbe2-6e02-4d98-9e50-94f772d89ac0] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 860.889039] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 860.889946] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-595fa78a-4c89-464f-900f-4bdace370596 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.899682] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 860.899682] env[66641]: value = "task-5146187" [ 860.899682] env[66641]: _type = "Task" [ 860.899682] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.910958] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146187, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.168069] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 861.169120] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e28212e-78a1-4d64-9b42-ee5e4fe7f0a2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.178942] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 861.179154] env[66641]: ERROR oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk due to incomplete transfer. [ 861.179402] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9e6e24e3-7540-46e5-9aa3-34e0f9f523a7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.187600] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8f96124-f302-45ef-8d45-9437c46bc84c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.193311] env[66641]: DEBUG oslo_vmware.rw_handles [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5286bd28-c00a-051d-f0ae-28ac9b1014b7/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 861.193530] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Uploaded image ccac96b4-e93b-428d-b8e5-d64b8daec92a to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 861.195978] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 861.196659] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-ba95c06c-4bf3-423d-9966-6f9c50e03e6c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.201562] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75ba0eb-74f9-4d02-ab0a-ee46c0f720eb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.205919] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 861.205919] env[66641]: value = "task-5146188" [ 861.205919] env[66641]: _type = "Task" [ 861.205919] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.235694] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82870837-996f-4499-9f56-0297f49912c7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.241569] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146188, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.247354] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-148ec411-0242-4410-ba42-328dc0c6a1e1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.261633] env[66641]: DEBUG nova.compute.provider_tree [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.350063] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 22899269-28fc-446e-9533-1f6862a9e8db] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 861.410483] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146187, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.717480] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146188, 'name': Destroy_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.765063] env[66641]: DEBUG nova.scheduler.client.report [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 861.853310] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: fb1bf80c-b243-4c25-9aaf-c459ca926090] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 861.910627] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146187, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.220408] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146188, 'name': Destroy_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.271048] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 862.271649] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 862.358052] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: d14528cb-f26e-4fe5-8bf9-34900571fe03] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 862.411669] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146187, 'name': CloneVM_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.725759] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146188, 'name': Destroy_Task, 'duration_secs': 1.400561} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.726608] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroyed the VM [ 862.726608] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 862.726738] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9225cc78-68e8-4ec9-9f83-1e43e60aa1c0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.735056] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 862.735056] env[66641]: value = "task-5146189" [ 862.735056] env[66641]: _type = "Task" [ 862.735056] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.746153] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146189, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.776996] env[66641]: DEBUG nova.compute.utils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 862.778416] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Allocating IP information in the background. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2019}} [ 862.779033] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] allocate_for_instance() {{(pid=66641) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1212}} [ 862.779033] env[66641]: WARNING neutronclient.v2_0.client [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.779276] env[66641]: WARNING neutronclient.v2_0.client [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] The python binding code in neutronclient is deprecated in favor of OpenstackSDK, please use that as this will be removed in a future release. [ 862.779886] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 862.780683] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 862.790084] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 862.830369] env[66641]: DEBUG nova.policy [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4dcb7f4435ed4cd69dcfc60c6da1671c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6b454fd430514418977cdd9276d504a1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=66641) authorize /opt/stack/nova/nova/policy.py:192}} [ 862.861810] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 434e66b1-5ea7-43d2-9b41-bdeda2a312d7] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 862.913973] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146187, 'name': CloneVM_Task, 'duration_secs': 1.573115} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.913973] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Created linked-clone VM from snapshot [ 862.914270] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5196b98-4f91-43a8-a3c2-700d0883c407 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.922937] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Uploading image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 862.953649] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 862.953649] env[66641]: value = "vm-1000724" [ 862.953649] env[66641]: _type = "VirtualMachine" [ 862.953649] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 862.953996] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-8b87d5f8-08e0-4256-b4ac-0610bab0cfc2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.962138] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lease: (returnval){ [ 862.962138] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3c485-8a89-854a-2021-16934678176b" [ 862.962138] env[66641]: _type = "HttpNfcLease" [ 862.962138] env[66641]: } obtained for exporting VM: (result){ [ 862.962138] env[66641]: value = "vm-1000724" [ 862.962138] env[66641]: _type = "VirtualMachine" [ 862.962138] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 862.962602] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the lease: (returnval){ [ 862.962602] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3c485-8a89-854a-2021-16934678176b" [ 862.962602] env[66641]: _type = "HttpNfcLease" [ 862.962602] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 862.970282] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 862.970282] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3c485-8a89-854a-2021-16934678176b" [ 862.970282] env[66641]: _type = "HttpNfcLease" [ 862.970282] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 863.078948] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Successfully created port: 620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:579}} [ 863.248693] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146189, 'name': RemoveSnapshot_Task} progress is 58%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.366390] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 695d1824-cf34-46c1-b623-b474d2f1f8ef] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 863.473115] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 863.473115] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3c485-8a89-854a-2021-16934678176b" [ 863.473115] env[66641]: _type = "HttpNfcLease" [ 863.473115] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 863.473454] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 863.473454] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b3c485-8a89-854a-2021-16934678176b" [ 863.473454] env[66641]: _type = "HttpNfcLease" [ 863.473454] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 863.474265] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa0e504-a13e-4977-8890-ec1fc03718b1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.482781] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 863.482961] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 863.574250] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-32d4c53f-18d1-43a9-b4c5-99b4f0d589f8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.748259] env[66641]: DEBUG oslo_vmware.api [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146189, 'name': RemoveSnapshot_Task, 'duration_secs': 0.56319} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.748259] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 863.748732] env[66641]: INFO nova.compute.manager [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 16.31 seconds to snapshot the instance on the hypervisor. [ 863.802508] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 863.827807] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 863.828664] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 863.828664] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 863.828982] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 863.829213] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 863.830175] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 863.831734] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.832028] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 863.832392] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 863.832502] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 863.832796] env[66641]: DEBUG nova.virt.hardware [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 863.834777] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c263d9d-75b8-48a2-92e9-7253b2313717 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.846033] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51baa20a-b84a-4256-9aee-0d1645211707 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.871395] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: e7bfb26e-b326-4a39-8f20-79dbbfb74eb8] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 864.297334] env[66641]: DEBUG nova.compute.manager [None req-7c37aee8-0742-495e-a16d-8e37276bb9a5 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Found 1 images (rotation: 2) {{(pid=66641) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 864.379064] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 5963a8ba-9d2c-42c3-a5da-25c29bf9e763] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 864.616443] env[66641]: DEBUG nova.compute.manager [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Received event network-vif-plugged-620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 864.616945] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] Acquiring lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 864.617251] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 864.617543] env[66641]: DEBUG oslo_concurrency.lockutils [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 864.617724] env[66641]: DEBUG nova.compute.manager [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] No waiting events found dispatching network-vif-plugged-620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 864.618389] env[66641]: WARNING nova.compute.manager [req-bd3b5a49-6c24-49e8-b3c9-ad186793759f req-88affdbb-bf2d-4ad0-90d4-ca919d72102b service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Received unexpected event network-vif-plugged-620784b2-23a1-4c14-9c20-331112a41c6c for instance with vm_state building and task_state spawning. [ 864.662593] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Successfully updated port: 620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) _update_port /opt/stack/nova/nova/network/neutron.py:607}} [ 864.883103] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 6b19204d-032c-410b-bb84-1af23d142edc] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 865.170340] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.170340] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquired lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 865.170340] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 865.387041] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: fe143d94-57ce-46d4-85e3-80d84af66dfc] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 865.673566] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.673982] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.679231] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 865.742853] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.743911] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.822553] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 865.823036] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 865.891085] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: d1e5ad0b-1e3c-436e-8cf5-136cbcf6f4cc] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 865.914836] env[66641]: DEBUG nova.network.neutron [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Updating instance_info_cache with network_info: [{"id": "620784b2-23a1-4c14-9c20-331112a41c6c", "address": "fa:16:3e:e1:ee:df", "network": {"id": "3824ad57-8d0c-413f-9a1d-b96ddb38153f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1281686137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b454fd430514418977cdd9276d504a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33fdc099-7497-41c1-b40c-1558937132d4", "external-id": "nsx-vlan-transportzone-764", "segmentation_id": 764, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620784b2-23", "ovs_interfaceid": "620784b2-23a1-4c14-9c20-331112a41c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 866.190870] env[66641]: DEBUG nova.compute.manager [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 866.192136] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0587b2-09ca-4783-9322-6ae3c6e6a77f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.395207] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 47aaed25-542b-4ceb-9adf-6a4953c8c95d] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 866.417977] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Releasing lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 866.418235] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Instance network_info: |[{"id": "620784b2-23a1-4c14-9c20-331112a41c6c", "address": "fa:16:3e:e1:ee:df", "network": {"id": "3824ad57-8d0c-413f-9a1d-b96ddb38153f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1281686137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b454fd430514418977cdd9276d504a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33fdc099-7497-41c1-b40c-1558937132d4", "external-id": "nsx-vlan-transportzone-764", "segmentation_id": 764, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620784b2-23", "ovs_interfaceid": "620784b2-23a1-4c14-9c20-331112a41c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2034}} [ 866.419463] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:ee:df', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '33fdc099-7497-41c1-b40c-1558937132d4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '620784b2-23a1-4c14-9c20-331112a41c6c', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.428163] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Creating folder: Project (6b454fd430514418977cdd9276d504a1). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 866.428866] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-815a53eb-a6bd-481e-9f20-9ba49b068125 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.442101] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Created folder: Project (6b454fd430514418977cdd9276d504a1) in parent group-v1000566. [ 866.442407] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Creating folder: Instances. Parent ref: group-v1000725. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 866.442685] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2abd35ab-8b80-40f0-aee6-d113e62ea8a5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.456071] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Created folder: Instances in parent group-v1000725. [ 866.456370] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 866.456680] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.456896] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ffac5d83-7a94-4073-a855-88fa53f4ff70 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.479311] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.479311] env[66641]: value = "task-5146195" [ 866.479311] env[66641]: _type = "Task" [ 866.479311] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.488209] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146195, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.652245] env[66641]: DEBUG nova.compute.manager [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Received event network-changed-620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 866.652612] env[66641]: DEBUG nova.compute.manager [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Refreshing instance network info cache due to event network-changed-620784b2-23a1-4c14-9c20-331112a41c6c. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 866.652866] env[66641]: DEBUG oslo_concurrency.lockutils [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Acquiring lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.653025] env[66641]: DEBUG oslo_concurrency.lockutils [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Acquired lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.653189] env[66641]: DEBUG nova.network.neutron [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Refreshing network info cache for port 620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 866.706291] env[66641]: INFO nova.compute.manager [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] instance snapshotting [ 866.706291] env[66641]: DEBUG nova.objects.instance [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lazy-loading 'flavor' on Instance uuid c95b481a-7956-410d-971c-7d94911230bb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 866.899419] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: dfa8c73b-db57-42a9-a9a4-cf812f5b2949] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 866.990539] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146195, 'name': CreateVM_Task, 'duration_secs': 0.358995} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.990753] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 866.991549] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 866.991911] env[66641]: WARNING openstack [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 866.997045] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.997213] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 866.997552] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 866.997849] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ff44385-b692-4b97-aaa6-93f7b642e4cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.003499] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 867.003499] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524302d1-645d-5144-8ef1-22ab5caced27" [ 867.003499] env[66641]: _type = "Task" [ 867.003499] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.012703] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524302d1-645d-5144-8ef1-22ab5caced27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.156711] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.157159] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.215167] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41e3009-fb7a-413d-bfdf-fcfc9e416344 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.237953] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09388877-d6b6-4fd9-99e4-b60066000d78 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.283428] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.283801] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.368265] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 867.368687] env[66641]: WARNING openstack [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 867.403142] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: d972a7e7-a04d-4244-a5b5-4aa5d57bf1e3] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 867.457261] env[66641]: DEBUG nova.network.neutron [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Updated VIF entry in instance network info cache for port 620784b2-23a1-4c14-9c20-331112a41c6c. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 867.457611] env[66641]: DEBUG nova.network.neutron [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Updating instance_info_cache with network_info: [{"id": "620784b2-23a1-4c14-9c20-331112a41c6c", "address": "fa:16:3e:e1:ee:df", "network": {"id": "3824ad57-8d0c-413f-9a1d-b96ddb38153f", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1281686137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6b454fd430514418977cdd9276d504a1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "33fdc099-7497-41c1-b40c-1558937132d4", "external-id": "nsx-vlan-transportzone-764", "segmentation_id": 764, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap620784b2-23", "ovs_interfaceid": "620784b2-23a1-4c14-9c20-331112a41c6c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 867.515375] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]524302d1-645d-5144-8ef1-22ab5caced27, 'name': SearchDatastore_Task, 'duration_secs': 0.015237} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.515706] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 867.515946] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.516219] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.516359] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 867.516538] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.516838] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5099a6e7-8464-45f0-9e46-ce216a085503 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.527099] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.527333] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.528474] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ef6487a-97b9-46ff-bae6-1cc40158b7f8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.534750] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 867.534750] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526d2546-7eb0-7587-9ea9-11aca48b1e08" [ 867.534750] env[66641]: _type = "Task" [ 867.534750] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.543067] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526d2546-7eb0-7587-9ea9-11aca48b1e08, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.750760] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 867.751138] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-955ce012-c755-4d33-9bf8-389d49821d57 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.761900] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 867.761900] env[66641]: value = "task-5146196" [ 867.761900] env[66641]: _type = "Task" [ 867.761900] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.773298] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146196, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.907116] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 420544a8-1a02-42d5-8a9a-e3e8b6a11a0c] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 867.960635] env[66641]: DEBUG oslo_concurrency.lockutils [req-eb2e6d25-88f2-4d99-acf5-d829ce51e3ac req-0b447b8d-d58c-433b-86d8-7fe6c1dbba17 service nova] Releasing lock "refresh_cache-27a9ee4a-7042-4d7a-b62a-05c25a84caef" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.046587] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526d2546-7eb0-7587-9ea9-11aca48b1e08, 'name': SearchDatastore_Task, 'duration_secs': 0.010323} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.047610] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-434a63a6-0091-46c2-a161-9b4ad26f564d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.054307] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 868.054307] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fedee9-0d36-5c17-c45e-8b98d4024a5e" [ 868.054307] env[66641]: _type = "Task" [ 868.054307] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.063209] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fedee9-0d36-5c17-c45e-8b98d4024a5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.274563] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146196, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.410783] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 4ebc0f2b-bb79-4941-b8f2-082560c80f8b] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 868.568023] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52fedee9-0d36-5c17-c45e-8b98d4024a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.010685} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.568023] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 868.568023] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 27a9ee4a-7042-4d7a-b62a-05c25a84caef/27a9ee4a-7042-4d7a-b62a-05c25a84caef.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.568023] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-eddccece-8dca-4c9b-a9c7-abef4d1ff750 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.574876] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 868.574876] env[66641]: value = "task-5146197" [ 868.574876] env[66641]: _type = "Task" [ 868.574876] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.583951] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146197, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.775059] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146196, 'name': CreateSnapshot_Task, 'duration_secs': 0.516958} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.776543] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 868.776771] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8046cf8-278d-4f5e-964e-1d1f9197dc95 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.914565] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: ccf46c5c-1c79-4672-ad9f-ea61042097d5] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 869.086589] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146197, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.296942] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 869.297530] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a0498572-e019-43aa-ad96-ab026aeefe23 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.307094] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 869.307094] env[66641]: value = "task-5146198" [ 869.307094] env[66641]: _type = "Task" [ 869.307094] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.316283] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146198, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.418259] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] [instance: 25203771-cf4d-4fc1-a1bf-f2a0200296b7] Instance has had 0 of 5 cleanup attempts {{(pid=66641) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11945}} [ 869.588876] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146197, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.521482} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.589193] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] 27a9ee4a-7042-4d7a-b62a-05c25a84caef/27a9ee4a-7042-4d7a-b62a-05c25a84caef.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 869.589406] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 869.589674] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b9f8b6c6-0ae8-43ba-82dc-18cdfe6e9de3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.598856] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 869.598856] env[66641]: value = "task-5146199" [ 869.598856] env[66641]: _type = "Task" [ 869.598856] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.610599] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146199, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.819302] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146198, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.922280] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 869.922477] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Cleaning up deleted instances with incomplete migration {{(pid=66641) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11970}} [ 870.109747] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146199, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.110552} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.109747] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.110398] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06e31d7-c7b7-4246-ab8a-ddc98d3e14cd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.133386] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Reconfiguring VM instance instance-00000031 to attach disk [datastore2] 27a9ee4a-7042-4d7a-b62a-05c25a84caef/27a9ee4a-7042-4d7a-b62a-05c25a84caef.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.133711] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c93093f-9c0b-4e26-a5da-45c0c4434d62 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.156025] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 870.156025] env[66641]: value = "task-5146200" [ 870.156025] env[66641]: _type = "Task" [ 870.156025] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.163491] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146200, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.318974] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146198, 'name': CloneVM_Task} progress is 95%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.425037] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 870.664934] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146200, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.821120] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146198, 'name': CloneVM_Task, 'duration_secs': 1.129699} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.821402] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created linked-clone VM from snapshot [ 870.822262] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1f629a-7f53-4e22-a9da-c44ad3b3dd1f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.830363] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Uploading image 5e2e348e-7ec1-4d7d-96bf-3329ea0e3182 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 870.853850] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 870.853850] env[66641]: value = "vm-1000729" [ 870.853850] env[66641]: _type = "VirtualMachine" [ 870.853850] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 870.854168] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-cca0128b-36bc-4698-ba22-5cde1b394ebb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.861933] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease: (returnval){ [ 870.861933] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aeb957-45ef-c91f-1f61-795a822ab6c2" [ 870.861933] env[66641]: _type = "HttpNfcLease" [ 870.861933] env[66641]: } obtained for exporting VM: (result){ [ 870.861933] env[66641]: value = "vm-1000729" [ 870.861933] env[66641]: _type = "VirtualMachine" [ 870.861933] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 870.862314] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the lease: (returnval){ [ 870.862314] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aeb957-45ef-c91f-1f61-795a822ab6c2" [ 870.862314] env[66641]: _type = "HttpNfcLease" [ 870.862314] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 870.869193] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 870.869193] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aeb957-45ef-c91f-1f61-795a822ab6c2" [ 870.869193] env[66641]: _type = "HttpNfcLease" [ 870.869193] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 870.997028] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.997998] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034857a0-17dc-46c7-8c42-1eda90d56851 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.004299] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 871.004454] env[66641]: ERROR oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk due to incomplete transfer. [ 871.004700] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-fcae42bf-9401-447c-b6e3-9c1d8c162001 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.012124] env[66641]: DEBUG oslo_vmware.rw_handles [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52b14dcf-f561-0908-cd82-eada9d25bb7d/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 871.012317] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Uploaded image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 871.014884] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 871.015146] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2f02e0ef-c840-468a-bc4a-1f21a22e3574 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.020841] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 871.020841] env[66641]: value = "task-5146202" [ 871.020841] env[66641]: _type = "Task" [ 871.020841] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.029489] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146202, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.165375] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146200, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.373258] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 871.373258] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aeb957-45ef-c91f-1f61-795a822ab6c2" [ 871.373258] env[66641]: _type = "HttpNfcLease" [ 871.373258] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 871.373258] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 871.373258] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52aeb957-45ef-c91f-1f61-795a822ab6c2" [ 871.373258] env[66641]: _type = "HttpNfcLease" [ 871.373258] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 871.373698] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65107425-4be9-4a31-b9ce-2b3ffb15d199 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.381252] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 871.381420] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 871.476950] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-410063b3-e726-4e8a-99bb-232f5aef9378 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.530724] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146202, 'name': Destroy_Task, 'duration_secs': 0.346611} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.530937] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Destroyed the VM [ 871.531184] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 871.531442] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-75c2ef6f-ded9-4dbe-9352-fee56cd6082e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.537560] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 871.537560] env[66641]: value = "task-5146203" [ 871.537560] env[66641]: _type = "Task" [ 871.537560] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.545526] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146203, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.667437] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146200, 'name': ReconfigVM_Task, 'duration_secs': 1.36835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.667779] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Reconfigured VM instance instance-00000031 to attach disk [datastore2] 27a9ee4a-7042-4d7a-b62a-05c25a84caef/27a9ee4a-7042-4d7a-b62a-05c25a84caef.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.668496] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc4d3d2e-1397-4c3b-b4e6-99533032e554 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.676326] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 871.676326] env[66641]: value = "task-5146204" [ 871.676326] env[66641]: _type = "Task" [ 871.676326] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.685382] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146204, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.048499] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146203, 'name': RemoveSnapshot_Task, 'duration_secs': 0.41708} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.048859] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 872.049196] env[66641]: DEBUG nova.compute.manager [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 872.050047] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3668adf4-fde5-4101-87d4-cfd0345288b8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.187093] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146204, 'name': Rename_Task, 'duration_secs': 0.263508} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.187501] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.187904] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02807b93-bd43-44b0-adea-080f85054f7f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.195697] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 872.195697] env[66641]: value = "task-5146205" [ 872.195697] env[66641]: _type = "Task" [ 872.195697] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.205731] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.564144] env[66641]: INFO nova.compute.manager [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Shelve offloading [ 872.708953] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146205, 'name': PowerOnVM_Task} progress is 88%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.068134] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.068488] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-682e7b2c-7984-4010-a0ec-c006ed9091c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.077062] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 873.077062] env[66641]: value = "task-5146206" [ 873.077062] env[66641]: _type = "Task" [ 873.077062] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.086829] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] VM already powered off {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 873.087267] env[66641]: DEBUG nova.compute.manager [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 873.088125] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0c1707-b749-44f4-b36f-935994e93af7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.095174] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.095333] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 873.095497] env[66641]: DEBUG nova.network.neutron [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 873.207404] env[66641]: DEBUG oslo_vmware.api [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146205, 'name': PowerOnVM_Task, 'duration_secs': 0.684498} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.207772] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.208103] env[66641]: INFO nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Took 9.41 seconds to spawn the instance on the hypervisor. [ 873.208343] env[66641]: DEBUG nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 873.209448] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b166f39-5a1e-4196-a642-148250bdde4e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.598547] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.598880] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.729787] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.730787] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.738959] env[66641]: INFO nova.compute.manager [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Took 13.74 seconds to build instance. [ 873.798024] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 873.798410] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 873.881699] env[66641]: DEBUG nova.network.neutron [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 873.925678] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.925889] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926052] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926200] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926342] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926694] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926694] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 873.926829] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}} [ 873.926882] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 874.240777] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4fbaa49a-79b2-492b-b61f-6ce2c98c9241 tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.257s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.385196] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 874.386032] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.386522] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.392216] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 874.392368] env[66641]: WARNING openstack [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 874.429216] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.429583] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.429705] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.429791] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 874.431179] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ce4f04-eda5-4121-bade-e7d61e232536 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.445869] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b11668f-fae7-4773-ab74-852740255865 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.460286] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f474643-fab1-4cbf-ad7c-bd785048a96a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.468322] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b0cf4ee-8d6b-4963-a70c-2cc908893d54 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.501400] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179139MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 874.501566] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.501773] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.672976] env[66641]: DEBUG nova.compute.manager [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-vif-unplugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 874.673375] env[66641]: DEBUG oslo_concurrency.lockutils [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.673612] env[66641]: DEBUG oslo_concurrency.lockutils [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.673828] env[66641]: DEBUG oslo_concurrency.lockutils [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.673971] env[66641]: DEBUG nova.compute.manager [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] No waiting events found dispatching network-vif-unplugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 874.674173] env[66641]: WARNING nova.compute.manager [req-83877e43-ffca-417c-9cb9-52e28520e1a9 req-79c90bfa-ea4c-4505-a189-4f4c9eb53a2c service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received unexpected event network-vif-unplugged-8bed634b-5119-478e-a71f-a832ef16f035 for instance with vm_state shelved and task_state shelving_offloading. [ 874.822515] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 874.823775] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-728a62e0-6088-4ff3-8750-0de52ae963f0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.832976] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 874.833136] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-aaf5be33-3237-4099-8ff3-e9a3163b9b9d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.903827] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.904095] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.904308] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleting the datastore file [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.904615] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-518f049c-036b-4944-a7b8-f9460a2bbe02 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.912898] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 874.912898] env[66641]: value = "task-5146208" [ 874.912898] env[66641]: _type = "Task" [ 874.912898] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.922958] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.923281] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.923519] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 874.923764] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 874.923937] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 874.925758] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146208, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.926343] env[66641]: INFO nova.compute.manager [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Terminating instance [ 875.423564] env[66641]: DEBUG oslo_vmware.api [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146208, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.154268} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.423845] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.424046] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.424228] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.430034] env[66641]: DEBUG nova.compute.manager [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 875.430249] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 875.431141] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fee169-9456-4c07-9d6b-d051882fab79 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.440053] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 875.440606] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-76689ca9-e3ab-4398-b919-b679af1f758f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.448804] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 875.448804] env[66641]: value = "task-5146209" [ 875.448804] env[66641]: _type = "Task" [ 875.448804] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.449951] env[66641]: INFO nova.scheduler.client.report [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleted allocations for instance cdfcd9d8-dbf7-4046-8338-42762fc389a1 [ 875.466337] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146209, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.532029] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance c95b481a-7956-410d-971c-7d94911230bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 875.532234] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 88ae00a2-6139-4258-b316-0f75032275ec actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 875.532393] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Instance 27a9ee4a-7042-4d7a-b62a-05c25a84caef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=66641) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1740}} [ 875.532591] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 875.532790] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=100GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] stats={'failed_builds': '0', 'num_instances': '4', 'num_vm_active': '3', 'num_task_image_uploading': '1', 'num_os_type_None': '4', 'num_proj_c48b059b6f7d49a8ba5ecf28125ddaa9': '1', 'io_workload': '0', 'num_vm_shelved': '1', 'num_task_shelving_offloading': '1', 'num_proj_f7e0123f065d40fe8b51e997ae3f6089': '1', 'num_task_None': '1', 'num_proj_3a9466d8e6fa49c59a14110b417664c6': '1', 'num_task_deleting': '1', 'num_proj_6b454fd430514418977cdd9276d504a1': '1'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 875.588154] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40ea21a9-a4da-4b0b-b522-83d36bd1316a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.597680] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e0d9d8-b16b-4db7-ac65-bc691a62f0f6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.632961] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6a1d80-c6ee-44af-9ce5-eac8b98190ec {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.643418] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e1dd42b-efe9-4bcd-984b-0a2de7062288 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.664743] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.959017] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 875.965074] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146209, 'name': PowerOffVM_Task, 'duration_secs': 0.223996} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.965349] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 875.965513] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 875.965779] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62fbd3a1-b143-43e9-a753-a1927e556890 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.035107] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 876.035453] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 876.035714] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Deleting the datastore file [datastore2] 27a9ee4a-7042-4d7a-b62a-05c25a84caef {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 876.036137] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-23436c74-6ad0-48ae-8d5e-fcf3844ecc15 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.046337] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for the task: (returnval){ [ 876.046337] env[66641]: value = "task-5146211" [ 876.046337] env[66641]: _type = "Task" [ 876.046337] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.057304] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146211, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.169104] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 876.557322] env[66641]: DEBUG oslo_vmware.api [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Task: {'id': task-5146211, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158835} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.557621] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 876.557822] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 876.558018] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 876.558190] env[66641]: INFO nova.compute.manager [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Took 1.13 seconds to destroy the instance on the hypervisor. [ 876.558444] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 876.558741] env[66641]: DEBUG nova.compute.manager [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 876.558905] env[66641]: DEBUG nova.network.neutron [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 876.559466] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.559719] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.606945] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 876.607261] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 876.677064] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 876.677258] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.175s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 876.677470] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.719s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 876.677706] env[66641]: DEBUG nova.objects.instance [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lazy-loading 'resources' on Instance uuid cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.727931] env[66641]: DEBUG nova.compute.manager [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-changed-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 876.728482] env[66641]: DEBUG nova.compute.manager [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing instance network info cache due to event network-changed-8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 876.729324] env[66641]: DEBUG oslo_concurrency.lockutils [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.729324] env[66641]: DEBUG oslo_concurrency.lockutils [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 876.729324] env[66641]: DEBUG nova.network.neutron [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 877.180404] env[66641]: DEBUG nova.objects.instance [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lazy-loading 'numa_topology' on Instance uuid cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 877.232797] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.233276] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.339294] env[66641]: DEBUG nova.network.neutron [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.377370] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.377805] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.449227] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 877.449612] env[66641]: WARNING openstack [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 877.540440] env[66641]: DEBUG nova.network.neutron [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updated VIF entry in instance network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 877.540746] env[66641]: DEBUG nova.network.neutron [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap8bed634b-51", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 877.683044] env[66641]: DEBUG nova.objects.base [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=66641) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 877.749135] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c2deda-bdd1-4d08-9198-5783dd914c1d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.757906] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93dcee0d-a479-452b-845c-2883f10035d4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.789701] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5345c6a-5843-490b-9ade-8f8bbfd696b3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.798334] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886d4985-6674-44e7-99c0-869089cb4ffd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.813085] env[66641]: DEBUG nova.compute.provider_tree [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.841653] env[66641]: INFO nova.compute.manager [-] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Took 1.28 seconds to deallocate network for instance. [ 878.044155] env[66641]: DEBUG oslo_concurrency.lockutils [req-686a2692-2da9-4167-9a10-1231df332a5d req-c9483553-1f94-4045-b147-23b69f91ce50 service nova] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 878.316770] env[66641]: DEBUG nova.scheduler.client.report [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 878.351902] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.759985] env[66641]: DEBUG nova.compute.manager [req-21547b5c-a6a1-4ddf-827c-831528feba1e req-124b327f-79c2-4b2c-9faa-c128d1338d07 service nova] [instance: 27a9ee4a-7042-4d7a-b62a-05c25a84caef] Received event network-vif-deleted-620784b2-23a1-4c14-9c20-331112a41c6c {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 878.828296] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.151s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 878.831946] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 878.831946] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.480s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 878.832820] env[66641]: DEBUG nova.objects.instance [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lazy-loading 'resources' on Instance uuid 27a9ee4a-7042-4d7a-b62a-05c25a84caef {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.340730] env[66641]: DEBUG oslo_concurrency.lockutils [None req-c4a0227c-be7d-4f33-b6e1-ef23b254b598 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 22.071s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 879.342108] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 0.510s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 879.342339] env[66641]: INFO nova.compute.manager [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Unshelving [ 879.411399] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c850a713-4d3b-45a6-8990-f94945807ac8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.420480] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54389d62-9b1c-4c24-a1b5-1b010b95153a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.452118] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f7442a-570c-4581-bf27-e6162363e0c4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.460748] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd6af8ed-9dfb-456a-8538-ecb71b09a013 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.476310] env[66641]: DEBUG nova.compute.provider_tree [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.979996] env[66641]: DEBUG nova.scheduler.client.report [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 880.255063] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 880.256404] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d14682c-ab40-451e-a359-d30f31e12432 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.264300] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 880.264463] env[66641]: ERROR oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk due to incomplete transfer. [ 880.264701] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3cd8352b-fbaf-4f85-ad82-1ae1b924e247 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.276088] env[66641]: DEBUG oslo_vmware.rw_handles [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5289f6cd-5159-9306-2e16-2faad7876f8c/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 880.276301] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Uploaded image 5e2e348e-7ec1-4d7d-96bf-3329ea0e3182 to the Glance image server {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 880.278823] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 880.279129] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-cf1ddaff-d7a9-46a7-a1eb-cc775d120b53 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.285537] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 880.285537] env[66641]: value = "task-5146212" [ 880.285537] env[66641]: _type = "Task" [ 880.285537] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.295737] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146212, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.370330] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 880.485717] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.654s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 880.488240] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.118s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 880.488467] env[66641]: DEBUG nova.objects.instance [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lazy-loading 'pci_requests' on Instance uuid cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.504767] env[66641]: INFO nova.scheduler.client.report [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Deleted allocations for instance 27a9ee4a-7042-4d7a-b62a-05c25a84caef [ 880.795775] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146212, 'name': Destroy_Task, 'duration_secs': 0.356145} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.796055] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroyed the VM [ 880.796310] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 880.796582] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-df4161f5-16d5-4ff2-b6fb-6415a96c8fff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.803756] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 880.803756] env[66641]: value = "task-5146213" [ 880.803756] env[66641]: _type = "Task" [ 880.803756] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.811978] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146213, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.994575] env[66641]: DEBUG nova.objects.instance [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lazy-loading 'numa_topology' on Instance uuid cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.012232] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ee1f2a0a-bb5c-41a2-bdd3-0703e7aa417d tempest-ServerTagsTestJSON-1785039782 tempest-ServerTagsTestJSON-1785039782-project-member] Lock "27a9ee4a-7042-4d7a-b62a-05c25a84caef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.089s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 881.315463] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146213, 'name': RemoveSnapshot_Task} progress is 47%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.497519] env[66641]: INFO nova.compute.claims [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 881.815138] env[66641]: DEBUG oslo_vmware.api [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146213, 'name': RemoveSnapshot_Task, 'duration_secs': 0.5734} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.815409] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleted Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 881.815640] env[66641]: INFO nova.compute.manager [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 14.60 seconds to snapshot the instance on the hypervisor. [ 882.386178] env[66641]: DEBUG nova.compute.manager [None req-1597390b-6f87-43d3-ac7b-94568709b9db tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Found 2 images (rotation: 2) {{(pid=66641) _rotate_backups /opt/stack/nova/nova/compute/manager.py:5074}} [ 882.451317] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 882.451565] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 882.562379] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217c5cf9-cc53-4aa4-bf75-d0ad1291c7ce {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.570673] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68c3c9c2-befd-4645-9c71-d1a7ef1ea7ee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.601538] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f4c7075-54cf-46cb-9400-19e134dbd2ca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.609828] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd2af34-a41b-406b-a5e8-417d41da4e16 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.623843] env[66641]: DEBUG nova.compute.provider_tree [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.955108] env[66641]: DEBUG nova.compute.utils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 883.127060] env[66641]: DEBUG nova.scheduler.client.report [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 883.266934] env[66641]: DEBUG nova.compute.manager [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 883.268683] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-051f4aa5-933a-4561-88ac-33212f10fcd1 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.458220] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.631923] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.144s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 883.632636] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 883.633008] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 883.679769] env[66641]: INFO nova.network.neutron [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating port 8bed634b-5119-478e-a71f-a832ef16f035 with attributes {'binding_host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 883.779964] env[66641]: INFO nova.compute.manager [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] instance snapshotting [ 883.780587] env[66641]: DEBUG nova.objects.instance [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lazy-loading 'flavor' on Instance uuid c95b481a-7956-410d-971c-7d94911230bb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 884.097263] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.097475] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.286384] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e31a1a-6b70-474e-a410-2cc29ef5d1da {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.306120] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7975e16d-3fd1-4407-855b-3f5c40b7a215 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.522301] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 884.522571] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 884.522885] env[66641]: INFO nova.compute.manager [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Attaching volume 486d945f-6f92-4c40-8761-1524cc243da7 to /dev/sdb [ 884.554539] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c5ffdf7-e831-43ef-9faa-f095bd19774d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.561756] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a10265-1d31-4b89-ab32-98c0fa10b646 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.577652] env[66641]: DEBUG nova.virt.block_device [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating existing volume attachment record: 4305ae9a-26fe-4d99-a66d-890da0596c24 {{(pid=66641) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 884.600339] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 884.817532] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 884.817858] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-2e4a9b15-064f-4ec9-9b44-3c7193e3e936 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.824990] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 884.824990] env[66641]: value = "task-5146215" [ 884.824990] env[66641]: _type = "Task" [ 884.824990] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.833439] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146215, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.129732] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.130181] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.131993] env[66641]: INFO nova.compute.claims [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.150881] env[66641]: DEBUG nova.compute.manager [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 885.151134] env[66641]: DEBUG oslo_concurrency.lockutils [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 885.151375] env[66641]: DEBUG oslo_concurrency.lockutils [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 885.151815] env[66641]: DEBUG oslo_concurrency.lockutils [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 885.152161] env[66641]: DEBUG nova.compute.manager [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] No waiting events found dispatching network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) pop_instance_event /opt/stack/nova/nova/compute/manager.py:344}} [ 885.152508] env[66641]: WARNING nova.compute.manager [req-842f6f2f-c13c-4d2d-812b-790a3bbad3d8 req-3088fb01-4f3e-4ce4-96e3-0d51749e5dfd service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received unexpected event network-vif-plugged-8bed634b-5119-478e-a71f-a832ef16f035 for instance with vm_state shelved_offloaded and task_state spawning. [ 885.238261] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.238368] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 885.238544] env[66641]: DEBUG nova.network.neutron [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 885.335946] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146215, 'name': CreateSnapshot_Task, 'duration_secs': 0.460217} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.336319] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created Snapshot of the VM instance {{(pid=66641) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 885.337051] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44af3f2f-808a-4042-b94f-d1ad468ca675 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.741928] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.742371] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.857456] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Creating linked-clone VM from snapshot {{(pid=66641) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 885.861198] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.861453] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 885.867706] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-489808a0-bcec-437a-acc2-7a77b1e7c570 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.879064] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 885.879064] env[66641]: value = "task-5146218" [ 885.879064] env[66641]: _type = "Task" [ 885.879064] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.887457] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146218, 'name': CloneVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.935781] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 885.936659] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 886.024401] env[66641]: DEBUG nova.network.neutron [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 886.205536] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02532ec-a4fe-42bd-a48a-1ee6da8c0afd {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.213468] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9db416a2-f2f5-4472-9a26-c7d8049af051 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.245204] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65df0e61-f5d1-4094-819a-44592ff0b31e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.253784] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207e30b3-d09e-41b6-a534-1e58e3e06cdc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.267383] env[66641]: DEBUG nova.compute.provider_tree [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.389490] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146218, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.528183] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 886.560899] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='c4ca736574fbb77339ca8f9c83b37632',container_format='bare',created_at=2025-12-15T15:53:28Z,direct_url=,disk_format='vmdk',id=642937a6-6cd5-48bb-b438-d352bd1dc6f0,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-485403270-shelved',owner='f7e0123f065d40fe8b51e997ae3f6089',properties=ImageMetaProps,protected=,size=31671808,status='active',tags=,updated_at=2025-12-15T15:53:42Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 886.561182] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 886.561340] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 886.562582] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 886.562582] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 886.562582] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 886.562582] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.562582] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 886.562739] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 886.562739] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 886.562739] env[66641]: DEBUG nova.virt.hardware [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 886.563470] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-717573ab-b42a-4ded-bf77-a9c248a333fe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.572137] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7baa4a59-2fcf-41af-a909-30f597ca65ab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.586180] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:6f:9d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '271fe7a0-dfd7-409b-920a-cf83ef1a86a3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8bed634b-5119-478e-a71f-a832ef16f035', 'vif_model': 'vmxnet3'}] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.593562] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 886.593814] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 886.594050] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb8627b7-a026-4cf7-a2fd-da0da3cfa5f8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.615097] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.615097] env[66641]: value = "task-5146219" [ 886.615097] env[66641]: _type = "Task" [ 886.615097] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.624119] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146219, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.770853] env[66641]: DEBUG nova.scheduler.client.report [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 886.889854] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146218, 'name': CloneVM_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.125771] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146219, 'name': CreateVM_Task, 'duration_secs': 0.433798} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.125895] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 887.126765] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.127135] env[66641]: WARNING openstack [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.132135] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.132313] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.132697] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 887.132996] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4c8749b-eb7e-46b9-9d67-4d57ca2ff50e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.138630] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 887.138630] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527c5fd8-237b-c928-2f9c-4535df5e4d69" [ 887.138630] env[66641]: _type = "Task" [ 887.138630] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.153509] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527c5fd8-237b-c928-2f9c-4535df5e4d69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.187265] env[66641]: DEBUG nova.compute.manager [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-changed-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 887.187410] env[66641]: DEBUG nova.compute.manager [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing instance network info cache due to event network-changed-8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 887.187630] env[66641]: DEBUG oslo_concurrency.lockutils [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.187852] env[66641]: DEBUG oslo_concurrency.lockutils [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.187922] env[66641]: DEBUG nova.network.neutron [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 887.277371] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.147s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 887.277966] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Start building networks asynchronously for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2873}} [ 887.390553] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146218, 'name': CloneVM_Task, 'duration_secs': 1.446785} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.390838] env[66641]: INFO nova.virt.vmwareapi.vmops [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Created linked-clone VM from snapshot [ 887.391608] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76d75c8-2144-4ae0-9f2c-5afe5dd0bdfc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.399530] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Uploading image 090f8007-abf0-4b47-ab1f-9af2825bcf40 {{(pid=66641) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 887.422381] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.422381] env[66641]: value = "vm-1000733" [ 887.422381] env[66641]: _type = "VirtualMachine" [ 887.422381] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.422686] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-411386da-288f-470b-81a0-2f04725f7a76 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.430863] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease: (returnval){ [ 887.430863] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529c2239-60a1-e76a-9d82-101ab7d2cf7b" [ 887.430863] env[66641]: _type = "HttpNfcLease" [ 887.430863] env[66641]: } obtained for exporting VM: (result){ [ 887.430863] env[66641]: value = "vm-1000733" [ 887.430863] env[66641]: _type = "VirtualMachine" [ 887.430863] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.431224] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the lease: (returnval){ [ 887.431224] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529c2239-60a1-e76a-9d82-101ab7d2cf7b" [ 887.431224] env[66641]: _type = "HttpNfcLease" [ 887.431224] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.439704] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.439704] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529c2239-60a1-e76a-9d82-101ab7d2cf7b" [ 887.439704] env[66641]: _type = "HttpNfcLease" [ 887.439704] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.649327] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 887.649620] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Processing image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 887.649826] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.649966] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquired lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 887.650158] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.650418] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4105a087-ecfb-4181-a5dc-9df81e59ccbb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.659950] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.660135] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 887.660857] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb72fc81-4974-4844-b4ef-e5943d75cc5c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.666161] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 887.666161] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527bd40c-aab0-9cf5-2a1b-e790f6e31d86" [ 887.666161] env[66641]: _type = "Task" [ 887.666161] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.673773] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]527bd40c-aab0-9cf5-2a1b-e790f6e31d86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.690674] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.691020] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.783019] env[66641]: DEBUG nova.compute.utils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Using /dev/sd instead of None {{(pid=66641) get_next_device_name /opt/stack/nova/nova/compute/utils.py:239}} [ 887.785095] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Not allocating networking since 'none' was specified. {{(pid=66641) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:2015}} [ 887.813995] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.815038] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.873083] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 887.873466] env[66641]: WARNING openstack [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 887.939939] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.939939] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529c2239-60a1-e76a-9d82-101ab7d2cf7b" [ 887.939939] env[66641]: _type = "HttpNfcLease" [ 887.939939] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 887.940348] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 887.940348] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]529c2239-60a1-e76a-9d82-101ab7d2cf7b" [ 887.940348] env[66641]: _type = "HttpNfcLease" [ 887.940348] env[66641]: }. {{(pid=66641) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 887.940883] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffbdef32-65b5-4932-b10c-a123cfec7a63 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.948917] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 887.949100] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk for reading. {{(pid=66641) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 887.952207] env[66641]: DEBUG nova.network.neutron [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updated VIF entry in instance network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 887.952534] env[66641]: DEBUG nova.network.neutron [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.216", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 888.037095] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-5c0b92da-6c65-47f5-90f5-05f69709fa49 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.178948] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Preparing fetch location {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 888.179266] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Fetch image to [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226/OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226.vmdk {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 888.179455] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Downloading stream optimized image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 to [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226/OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226.vmdk on the data store datastore2 as vApp {{(pid=66641) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 888.179623] env[66641]: DEBUG nova.virt.vmwareapi.images [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Downloading image file data 642937a6-6cd5-48bb-b438-d352bd1dc6f0 to the ESX as VM named 'OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226' {{(pid=66641) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 888.261276] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 888.261276] env[66641]: value = "resgroup-9" [ 888.261276] env[66641]: _type = "ResourcePool" [ 888.261276] env[66641]: }. {{(pid=66641) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 888.261276] env[66641]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-4e0f4783-48e8-4f08-aa29-428f09b7a641 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.284030] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lease: (returnval){ [ 888.284030] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217c38c-01a7-2648-b9f7-fe8b438ebae6" [ 888.284030] env[66641]: _type = "HttpNfcLease" [ 888.284030] env[66641]: } obtained for vApp import into resource pool (val){ [ 888.284030] env[66641]: value = "resgroup-9" [ 888.284030] env[66641]: _type = "ResourcePool" [ 888.284030] env[66641]: }. {{(pid=66641) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 888.284309] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the lease: (returnval){ [ 888.284309] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217c38c-01a7-2648-b9f7-fe8b438ebae6" [ 888.284309] env[66641]: _type = "HttpNfcLease" [ 888.284309] env[66641]: } to be ready. {{(pid=66641) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 888.289943] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Start building block device mappings for instance. {{(pid=66641) _build_resources /opt/stack/nova/nova/compute/manager.py:2908}} [ 888.297028] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.297028] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217c38c-01a7-2648-b9f7-fe8b438ebae6" [ 888.297028] env[66641]: _type = "HttpNfcLease" [ 888.297028] env[66641]: } is initializing. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.507241] env[66641]: DEBUG oslo_concurrency.lockutils [req-b21a9ad7-ab07-4eef-bcdf-5ffe998de05e req-409c2b06-3e71-4ccb-84b9-4fef9f69a36a service nova] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 888.793135] env[66641]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.793135] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217c38c-01a7-2648-b9f7-fe8b438ebae6" [ 888.793135] env[66641]: _type = "HttpNfcLease" [ 888.793135] env[66641]: } is ready. {{(pid=66641) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 888.793574] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 888.793574] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]5217c38c-01a7-2648-b9f7-fe8b438ebae6" [ 888.793574] env[66641]: _type = "HttpNfcLease" [ 888.793574] env[66641]: }. {{(pid=66641) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 888.794427] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14023c39-47ce-48aa-b12f-6728c10a8e6a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.804686] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk from lease info. {{(pid=66641) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 888.805008] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating HTTP connection to write to file with size = 31671808 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk. {{(pid=66641) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 888.877352] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-229f4655-39d1-4a87-9306-acae3a7a8d07 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.307890] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Start spawning the instance on the hypervisor. {{(pid=66641) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2681}} [ 889.584398] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-12-15T15:47:15Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-12-15T15:46:54Z,direct_url=,disk_format='vmdk',id=cd910b37-6707-4868-b172-79fffc590a51,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='dde1b7d490614e5b8332835e29fc0c01',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-12-15T15:46:55Z,virtual_size=,visibility=), allow threads: False {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:571}} [ 889.584694] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Flavor limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:356}} [ 889.584940] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Image limits 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:360}} [ 889.585227] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Flavor pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:396}} [ 889.585827] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Image pref 0:0:0 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:400}} [ 889.585887] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=66641) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:438}} [ 889.586126] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.586368] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:479}} [ 889.586786] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Got 1 possible topologies {{(pid=66641) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:509}} [ 889.587041] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:583}} [ 889.587360] env[66641]: DEBUG nova.virt.hardware [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=66641) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:585}} [ 889.588359] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40937110-13b6-4604-ab9b-b60d1b2073cb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.603882] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f65f5f7-b0b3-4c63-95fa-c33cd4bce833 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.622606] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance VIF info [] {{(pid=66641) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.629575] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Creating folder: Project (53c78c0d2e954ca1b02658183eaffea3). Parent ref: group-v1000566. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.635043] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f8a8c8cd-1f2b-4b1f-aac1-f8715612d31f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.640644] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Volume attach. Driver type: vmdk {{(pid=66641) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 889.640907] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-1000731', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'name': 'volume-486d945f-6f92-4c40-8761-1524cc243da7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '88ae00a2-6139-4258-b316-0f75032275ec', 'attached_at': '', 'detached_at': '', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'serial': '486d945f-6f92-4c40-8761-1524cc243da7'} {{(pid=66641) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 889.641887] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db04bdfc-fa25-4624-b6f6-25108950e1f0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.668114] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9776d4e4-8e9f-4636-beaf-00fbca3e739b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.671366] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Created folder: Project (53c78c0d2e954ca1b02658183eaffea3) in parent group-v1000566. [ 889.671743] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Creating folder: Instances. Parent ref: group-v1000736. {{(pid=66641) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.672152] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4e9685db-7777-4a12-ad53-2c529dd47eb2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.701499] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] volume-486d945f-6f92-4c40-8761-1524cc243da7/volume-486d945f-6f92-4c40-8761-1524cc243da7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.710538] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-044e8396-41b0-4e58-8f4e-81548a474daf {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.726256] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Created folder: Instances in parent group-v1000736. [ 889.726647] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 889.731787] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Creating VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.732698] env[66641]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b468018-3566-44f8-a747-968bbdf7bda8 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.751663] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 889.751663] env[66641]: value = "task-5146225" [ 889.751663] env[66641]: _type = "Task" [ 889.751663] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.765567] env[66641]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.765567] env[66641]: value = "task-5146226" [ 889.765567] env[66641]: _type = "Task" [ 889.765567] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.765977] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.778333] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146226, 'name': CreateVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.088955] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Completed reading data from the image iterator. {{(pid=66641) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 890.089167] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 890.090119] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86ff4ef4-2fcc-42e4-bd2b-7a29d54df433 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.097922] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 890.098210] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 890.098419] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-783c78e5-16da-486b-beef-28c1d30afb06 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.262984] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146225, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.277937] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146226, 'name': CreateVM_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.329926] env[66641]: DEBUG oslo_vmware.rw_handles [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/522a710a-9809-7196-8870-37db4e72f12e/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 890.329926] env[66641]: INFO nova.virt.vmwareapi.images [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Downloaded image file data 642937a6-6cd5-48bb-b438-d352bd1dc6f0 [ 890.330331] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11de70c8-f018-44a9-b765-fe7b71910cd4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.347815] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e0116f3-9c37-471a-8b94-6ebbf3b849f2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.379906] env[66641]: INFO nova.virt.vmwareapi.images [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] The imported VM was unregistered [ 890.382063] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Caching image {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 890.382417] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Creating directory with path [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.382820] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-22eb8ff0-b457-401e-87ab-1111e5ec8756 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.395544] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Created directory with path [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0 {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.395782] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226/OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226.vmdk to [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk. {{(pid=66641) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 890.396022] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-3d058301-9194-449c-9473-5b984f84fa01 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.405219] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 890.405219] env[66641]: value = "task-5146228" [ 890.405219] env[66641]: _type = "Task" [ 890.405219] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.414870] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.766595] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146225, 'name': ReconfigVM_Task, 'duration_secs': 0.732217} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.766899] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfigured VM instance instance-0000002c to attach disk [datastore2] volume-486d945f-6f92-4c40-8761-1524cc243da7/volume-486d945f-6f92-4c40-8761-1524cc243da7.vmdk or device None with type thin {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.772580] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db78bd44-2468-4593-ad3a-a52bcb5691f4 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.795388] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146226, 'name': CreateVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.797268] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 890.797268] env[66641]: value = "task-5146229" [ 890.797268] env[66641]: _type = "Task" [ 890.797268] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.812027] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146229, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.917916] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.291558] env[66641]: DEBUG oslo_vmware.api [-] Task: {'id': task-5146226, 'name': CreateVM_Task, 'duration_secs': 1.144282} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.291911] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Created VM on the ESX host {{(pid=66641) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 891.292420] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.292583] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.293057] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:324}} [ 891.293368] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef665985-dc6b-464b-9cb2-c09e5a4368be {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.303016] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 891.303016] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b4a314-9bc0-26be-e7c9-a95e390458bb" [ 891.303016] env[66641]: _type = "Task" [ 891.303016] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.311065] env[66641]: DEBUG oslo_vmware.api [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146229, 'name': ReconfigVM_Task, 'duration_secs': 0.202943} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.311824] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-1000731', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'name': 'volume-486d945f-6f92-4c40-8761-1524cc243da7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '88ae00a2-6139-4258-b316-0f75032275ec', 'attached_at': '', 'detached_at': '', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'serial': '486d945f-6f92-4c40-8761-1524cc243da7'} {{(pid=66641) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 891.318395] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b4a314-9bc0-26be-e7c9-a95e390458bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.418928] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.821325] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52b4a314-9bc0-26be-e7c9-a95e390458bb, 'name': SearchDatastore_Task, 'duration_secs': 0.088772} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.821901] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 891.822178] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Processing image cd910b37-6707-4868-b172-79fffc590a51 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.822417] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.822563] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquired lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 891.822740] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.823211] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a01e46d-5d11-4552-88d1-932ef231f6e5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.844788] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=66641) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.844998] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=66641) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 891.846270] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b649c33-d907-4c94-9f61-44b9b70ec887 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.854154] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 891.854154] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bf07e5-c6f5-4992-887f-6dfbb351e1aa" [ 891.854154] env[66641]: _type = "Task" [ 891.854154] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.869710] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bf07e5-c6f5-4992-887f-6dfbb351e1aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.919828] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.362855] env[66641]: DEBUG nova.objects.instance [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lazy-loading 'flavor' on Instance uuid 88ae00a2-6139-4258-b316-0f75032275ec {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.371467] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]52bf07e5-c6f5-4992-887f-6dfbb351e1aa, 'name': SearchDatastore_Task, 'duration_secs': 0.090669} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.373244] env[66641]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc3f50af-b562-4541-b194-7e96c011d857 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.382385] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 892.382385] env[66641]: value = "session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526077b2-1e42-3c60-54a5-a65172267fdd" [ 892.382385] env[66641]: _type = "Task" [ 892.382385] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.398138] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526077b2-1e42-3c60-54a5-a65172267fdd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.424026] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.875817] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0bc8260d-be3a-4d7d-9525-e5e72068c3ca tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.353s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 892.894585] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': session[52cfd30a-797b-18a7-30e3-b8a2f0dfedea]526077b2-1e42-3c60-54a5-a65172267fdd, 'name': SearchDatastore_Task, 'duration_secs': 0.095498} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.895601] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Releasing lock "[datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 892.895854] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a8d805ee-1838-4e22-a2de-a48848b14f9c/a8d805ee-1838-4e22-a2de-a48848b14f9c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 892.896167] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d80ffa42-b2fc-41ec-b402-8f545d928add {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.904660] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 892.904660] env[66641]: value = "task-5146230" [ 892.904660] env[66641]: _type = "Task" [ 892.904660] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.917371] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146230, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.920694] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.420984] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146230, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.424940] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146228, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.59362} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.425371] env[66641]: INFO nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226/OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226.vmdk to [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk. [ 893.425444] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Cleaning up location [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226 {{(pid=66641) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 893.425594] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_4c5cf042-757b-4541-9d31-cc64b80f8226 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 893.425887] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04ae1966-1109-4e9f-8b02-6e340a86faef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.434767] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 893.434767] env[66641]: value = "task-5146231" [ 893.434767] env[66641]: _type = "Task" [ 893.434767] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.445576] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.916842] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146230, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.73586} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.917151] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/cd910b37-6707-4868-b172-79fffc590a51/cd910b37-6707-4868-b172-79fffc590a51.vmdk to [datastore2] a8d805ee-1838-4e22-a2de-a48848b14f9c/a8d805ee-1838-4e22-a2de-a48848b14f9c.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.917366] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Extending root virtual disk to 1048576 {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.917765] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-846ff4ad-7fd9-46ba-8da5-0d486d516a18 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.925672] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 893.925672] env[66641]: value = "task-5146232" [ 893.925672] env[66641]: _type = "Task" [ 893.925672] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.938157] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146232, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.948413] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190455} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.948693] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.948846] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Releasing lock "[datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 893.949102] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk to [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.949376] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ff6482de-61bb-409b-96a6-79cf845c2844 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.958241] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 893.958241] env[66641]: value = "task-5146233" [ 893.958241] env[66641]: _type = "Task" [ 893.958241] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.967879] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.076928] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 894.284332] env[66641]: DEBUG nova.compute.manager [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-changed-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 894.284577] env[66641]: DEBUG nova.compute.manager [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing instance network info cache due to event network-changed-8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 894.284791] env[66641]: DEBUG oslo_concurrency.lockutils [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Acquiring lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.284932] env[66641]: DEBUG oslo_concurrency.lockutils [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Acquired lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 894.285091] env[66641]: DEBUG nova.network.neutron [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Refreshing network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 894.441357] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146232, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080999} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.441735] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Extended root virtual disk {{(pid=66641) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 894.442800] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-963994ed-162e-4616-a347-1ddfadadc790 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.470393] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Reconfiguring VM instance instance-00000032 to attach disk [datastore2] a8d805ee-1838-4e22-a2de-a48848b14f9c/a8d805ee-1838-4e22-a2de-a48848b14f9c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 894.474381] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-91bf1d87-07dc-4a40-a05f-f9d4148b2805 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.498194] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.500301] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 894.500301] env[66641]: value = "task-5146234" [ 894.500301] env[66641]: _type = "Task" [ 894.500301] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.512725] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146234, 'name': ReconfigVM_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.788756] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.789285] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.931172] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 894.931710] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 894.969977] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.011501] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146234, 'name': ReconfigVM_Task} progress is 99%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.020215] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 895.020680] env[66641]: WARNING openstack [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 895.037469] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 895.037844] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 895.123069] env[66641]: DEBUG nova.network.neutron [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updated VIF entry in instance network info cache for port 8bed634b-5119-478e-a71f-a832ef16f035. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 895.123658] env[66641]: DEBUG nova.network.neutron [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [{"id": "8bed634b-5119-478e-a71f-a832ef16f035", "address": "fa:16:3e:6b:6f:9d", "network": {"id": "089f0304-3cbc-4d00-ba4f-18d6a67f473b", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-899271269-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f7e0123f065d40fe8b51e997ae3f6089", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "271fe7a0-dfd7-409b-920a-cf83ef1a86a3", "external-id": "nsx-vlan-transportzone-728", "segmentation_id": 728, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8bed634b-51", "ovs_interfaceid": "8bed634b-5119-478e-a71f-a832ef16f035", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 895.470684] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.512498] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146234, 'name': ReconfigVM_Task, 'duration_secs': 0.51859} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.512847] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Reconfigured VM instance instance-00000032 to attach disk [datastore2] a8d805ee-1838-4e22-a2de-a48848b14f9c/a8d805ee-1838-4e22-a2de-a48848b14f9c.vmdk or device None with type sparse {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 895.513662] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36079232-38ba-4dea-bda6-0500fcf25940 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.522879] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 895.522879] env[66641]: value = "task-5146235" [ 895.522879] env[66641]: _type = "Task" [ 895.522879] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.544961] env[66641]: INFO nova.compute.manager [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Detaching volume 486d945f-6f92-4c40-8761-1524cc243da7 [ 895.547459] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146235, 'name': Rename_Task} progress is 6%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.591901] env[66641]: INFO nova.virt.block_device [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Attempting to driver detach volume 486d945f-6f92-4c40-8761-1524cc243da7 from mountpoint /dev/sdb [ 895.592330] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Volume detach. Driver type: vmdk {{(pid=66641) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 895.592607] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-1000731', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'name': 'volume-486d945f-6f92-4c40-8761-1524cc243da7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '88ae00a2-6139-4258-b316-0f75032275ec', 'attached_at': '', 'detached_at': '', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'serial': '486d945f-6f92-4c40-8761-1524cc243da7'} {{(pid=66641) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 895.593625] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abcde84-0e76-4668-b7e7-eeea84fdebfe {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.618879] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af705f68-401d-496a-a397-0b1c51bc270d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.628022] env[66641]: DEBUG oslo_concurrency.lockutils [req-47493d58-442c-4552-aa85-0740b86a4758 req-187d6cac-8442-4d16-876e-a55476de0d47 service nova] Releasing lock "refresh_cache-cdfcd9d8-dbf7-4046-8338-42762fc389a1" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 895.629192] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17d1d42a-0848-4d1a-95de-fb4daf0c44e5 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.652200] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e961f93-116c-4b95-8115-83d28791c6f3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.669107] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] The volume has not been displaced from its original location: [datastore2] volume-486d945f-6f92-4c40-8761-1524cc243da7/volume-486d945f-6f92-4c40-8761-1524cc243da7.vmdk. No consolidation needed. {{(pid=66641) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 895.674733] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfiguring VM instance instance-0000002c to detach disk 2001 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 895.675176] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a53c1f74-282c-4756-a933-ded2088c57c3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.698614] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 895.698614] env[66641]: value = "task-5146236" [ 895.698614] env[66641]: _type = "Task" [ 895.698614] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.711468] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146236, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.973189] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.033520] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146235, 'name': Rename_Task, 'duration_secs': 0.203832} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.033795] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.034075] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c385dbd2-3c86-4d2f-b117-d50a7480e9ab {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.042786] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 896.042786] env[66641]: value = "task-5146237" [ 896.042786] env[66641]: _type = "Task" [ 896.042786] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.052122] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.209162] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146236, 'name': ReconfigVM_Task, 'duration_secs': 0.286013} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.209500] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Reconfigured VM instance instance-0000002c to detach disk 2001 {{(pid=66641) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 896.214821] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-35d79b17-569f-41b3-8987-2b43a41a99c9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.231861] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 896.231861] env[66641]: value = "task-5146238" [ 896.231861] env[66641]: _type = "Task" [ 896.231861] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.241613] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146238, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.472824] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.555151] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146237, 'name': PowerOnVM_Task} progress is 89%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.742176] env[66641]: DEBUG oslo_vmware.api [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146238, 'name': ReconfigVM_Task, 'duration_secs': 0.183856} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.742502] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-1000731', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'name': 'volume-486d945f-6f92-4c40-8761-1524cc243da7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False, 'enforce_multipath': False}, 'status': 'reserved', 'instance': '88ae00a2-6139-4258-b316-0f75032275ec', 'attached_at': '', 'detached_at': '', 'volume_id': '486d945f-6f92-4c40-8761-1524cc243da7', 'serial': '486d945f-6f92-4c40-8761-1524cc243da7'} {{(pid=66641) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 896.972172] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146233, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.638299} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.972437] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/642937a6-6cd5-48bb-b438-d352bd1dc6f0/642937a6-6cd5-48bb-b438-d352bd1dc6f0.vmdk to [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk {{(pid=66641) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 896.973291] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a6cc2a1-50ec-40a2-a6b6-a3a60bf3feb7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.995398] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk or device None with type streamOptimized {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.995716] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eeaa763-b26b-4d7f-bff2-1760ae837972 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.015697] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 897.015697] env[66641]: value = "task-5146239" [ 897.015697] env[66641]: _type = "Task" [ 897.015697] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.023941] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.057694] env[66641]: DEBUG oslo_vmware.api [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146237, 'name': PowerOnVM_Task, 'duration_secs': 0.680244} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.057968] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.058188] env[66641]: INFO nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Took 7.75 seconds to spawn the instance on the hypervisor. [ 897.058363] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 897.059256] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c6701f-c4fd-4597-900f-1568d38c9ea3 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.181227] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 897.298789] env[66641]: DEBUG nova.objects.instance [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lazy-loading 'flavor' on Instance uuid 88ae00a2-6139-4258-b316-0f75032275ec {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.528727] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.571212] env[66641]: DEBUG nova.compute.utils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Conflict updating instance a8d805ee-1838-4e22-a2de-a48848b14f9c. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} {{(pid=66641) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 897.574704] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance disappeared during build. {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2556}} [ 897.574875] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Unplugging VIFs for instance {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3050}} [ 897.575105] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.575280] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquired lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 897.575469] env[66641]: DEBUG nova.network.neutron [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 898.028594] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146239, 'name': ReconfigVM_Task, 'duration_secs': 0.938294} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.029204] env[66641]: DEBUG nova.virt.vmwareapi.volumeops [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Reconfigured VM instance instance-0000002b to attach disk [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1/cdfcd9d8-dbf7-4046-8338-42762fc389a1.vmdk or device None with type streamOptimized {{(pid=66641) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.029977] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6fb187ae-bb40-4a3d-977a-9ddc835e7efa {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.038403] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 898.038403] env[66641]: value = "task-5146240" [ 898.038403] env[66641]: _type = "Task" [ 898.038403] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.047807] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146240, 'name': Rename_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.078327] env[66641]: WARNING openstack [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 898.078779] env[66641]: WARNING openstack [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 898.084445] env[66641]: DEBUG nova.network.neutron [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 898.104539] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.104875] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.105140] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 898.105374] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 898.105587] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.108168] env[66641]: INFO nova.compute.manager [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Terminating instance [ 898.197670] env[66641]: DEBUG nova.network.neutron [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 898.309440] env[66641]: DEBUG oslo_concurrency.lockutils [None req-0b70d42d-bc5c-4f8a-a138-59705a8cf680 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.271s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 898.550211] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146240, 'name': Rename_Task, 'duration_secs': 0.200049} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.550569] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powering on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 898.550824] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-00528972-5d80-4d83-b356-2cce92c14dd9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.559757] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 898.559757] env[66641]: value = "task-5146241" [ 898.559757] env[66641]: _type = "Task" [ 898.559757] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.569718] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.612516] env[66641]: DEBUG nova.compute.manager [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 898.612702] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 898.613711] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d09dd1-1619-4b24-a583-9f098d2bf866 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.623710] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 898.624112] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-18d74057-d5c1-43cc-bd2c-8aab6464143b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.633119] env[66641]: DEBUG oslo_vmware.api [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 898.633119] env[66641]: value = "task-5146242" [ 898.633119] env[66641]: _type = "Task" [ 898.633119] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.645094] env[66641]: DEBUG oslo_vmware.api [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146242, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.672559] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 898.673644] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5427c4-9cf7-42eb-adc8-74bb83ed5fb2 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.681074] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk is in state: ready. {{(pid=66641) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 898.681318] env[66641]: ERROR oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk due to incomplete transfer.: glanceclient.exc.HTTPException: HTTP 410 Gone: Image 090f8007-abf0-4b47-ab1f-9af2825bcf40 could not be found after upload. The image may have been deleted during the upload, cleaning up the chunks uploaded. [ 898.681580] env[66641]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-8acdf124-8bb4-430f-b64b-aaced3ed44f6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.694575] env[66641]: DEBUG oslo_vmware.rw_handles [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529cce5b-f04b-32d5-2c6f-d7047bbaa282/disk-0.vmdk. {{(pid=66641) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 898.694772] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroying the VM {{(pid=66641) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 898.695072] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5db14429-4822-4a4f-a71f-3f03c0870dff {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.700723] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Releasing lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 898.700968] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3073}} [ 898.701220] env[66641]: DEBUG nova.compute.manager [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Skipping network deallocation for instance since networking was not requested. {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2323}} [ 898.706250] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 898.706250] env[66641]: value = "task-5146243" [ 898.706250] env[66641]: _type = "Task" [ 898.706250] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.718974] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146243, 'name': Destroy_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.070239] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146241, 'name': PowerOnVM_Task} progress is 78%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.143163] env[66641]: DEBUG oslo_vmware.api [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146242, 'name': PowerOffVM_Task, 'duration_secs': 0.257309} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.143542] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 899.143734] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 899.144014] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7f5c9ef1-69fb-4c63-a03a-d41ac719135b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.218179] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146243, 'name': Destroy_Task} progress is 33%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.256710] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 899.256961] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 899.257165] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Deleting the datastore file [datastore2] c95b481a-7956-410d-971c-7d94911230bb {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 899.257453] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-947d9471-bd78-4c3d-a9ec-382ea1f38a03 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.268462] env[66641]: DEBUG oslo_vmware.api [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for the task: (returnval){ [ 899.268462] env[66641]: value = "task-5146245" [ 899.268462] env[66641]: _type = "Task" [ 899.268462] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.280124] env[66641]: DEBUG oslo_vmware.api [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146245, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.413937] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.414221] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.414426] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "88ae00a2-6139-4258-b316-0f75032275ec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.414604] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.414768] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.417369] env[66641]: INFO nova.compute.manager [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Terminating instance [ 899.571133] env[66641]: DEBUG oslo_vmware.api [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146241, 'name': PowerOnVM_Task, 'duration_secs': 0.90879} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.571880] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powered on the VM {{(pid=66641) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 899.592394] env[66641]: WARNING nova.compute.manager [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Failed to delete snapshot from shelved instance (Image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 could not be found.).: nova.exception.ImageNotFound: Image 642937a6-6cd5-48bb-b438-d352bd1dc6f0 could not be found. [ 899.592774] env[66641]: DEBUG nova.compute.manager [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Checking state {{(pid=66641) _get_power_state /opt/stack/nova/nova/compute/manager.py:1828}} [ 899.593676] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9a790dd-6250-445c-8b21-4ba4b6c7a44f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.717851] env[66641]: DEBUG nova.compute.manager [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 899.717965] env[66641]: DEBUG nova.compute.manager [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing instance network info cache due to event network-changed-049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11840}} [ 899.718130] env[66641]: DEBUG oslo_concurrency.lockutils [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Acquiring lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 899.718313] env[66641]: DEBUG oslo_concurrency.lockutils [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Acquired lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 899.718557] env[66641]: DEBUG nova.network.neutron [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Refreshing network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2058}} [ 899.729410] env[66641]: DEBUG oslo_vmware.api [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Task: {'id': task-5146243, 'name': Destroy_Task, 'duration_secs': 0.679186} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.729687] env[66641]: INFO nova.virt.vmwareapi.vm_util [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroyed the VM [ 899.730221] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deleting Snapshot of the VM instance {{(pid=66641) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 899.730885] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bbb3eb44-5c2e-49fb-995c-2ce616c6ac71 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.738086] env[66641]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 899.738272] env[66641]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=66641) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 899.739057] env[66641]: DEBUG nova.compute.utils [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Cleaning up image 090f8007-abf0-4b47-ab1f-9af2825bcf40 {{(pid=66641) delete_image /opt/stack/nova/nova/compute/utils.py:1352}} [ 899.754344] env[66641]: INFO nova.scheduler.client.report [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Deleted allocations for instance a8d805ee-1838-4e22-a2de-a48848b14f9c [ 899.755031] env[66641]: DEBUG oslo_concurrency.lockutils [None req-da0b6159-9ca7-4f7a-a629-f3fa579116a6 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.657s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.755956] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.575s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.756378] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "a8d805ee-1838-4e22-a2de-a48848b14f9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 899.756677] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 899.756907] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 899.763897] env[66641]: INFO nova.compute.manager [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Terminating instance [ 899.783149] env[66641]: WARNING nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] In vmwareapi:vmops:_destroy_instance, exception while deleting the VM contents from the disk: oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore2] c95b481a-7956-410d-971c-7d94911230bb [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] Traceback (most recent call last): [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1113, in _destroy_instance [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] ds_util.file_delete(self._session, [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/nova/nova/virt/vmwareapi/ds_util.py", line 219, in file_delete [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] session._wait_for_task(file_delete_task) [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] return self.wait_for_task(task_ref) [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] return evt.wait() [ 899.783149] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] result = hub.switch() [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] return self.greenlet.switch() [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] self.f(*self.args, **self.kw) [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] raise exceptions.translate_fault(task_info.error) [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] oslo_vmware.exceptions.CannotDeleteFileException: Cannot delete file [datastore2] c95b481a-7956-410d-971c-7d94911230bb [ 899.783797] env[66641]: ERROR nova.virt.vmwareapi.vmops [instance: c95b481a-7956-410d-971c-7d94911230bb] [ 899.783797] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 899.784235] env[66641]: INFO nova.compute.manager [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 1.17 seconds to destroy the instance on the hypervisor. [ 899.784297] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 899.784559] env[66641]: DEBUG nova.compute.manager [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 899.784722] env[66641]: DEBUG nova.network.neutron [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 899.785574] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.786594] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.832187] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 899.832485] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 899.921696] env[66641]: DEBUG nova.compute.manager [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 899.921818] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.922713] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b2baac-5a9d-4568-92dc-137b8c7d5dca {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.931605] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.931993] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-56f439c5-d836-4766-9578-9e4c14191a44 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.939409] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 899.939409] env[66641]: value = "task-5146246" [ 899.939409] env[66641]: _type = "Task" [ 899.939409] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.949418] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146246, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.104682] env[66641]: DEBUG oslo_concurrency.lockutils [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.763s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.106879] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 6.030s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.107226] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 900.107469] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 900.107757] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 900.111668] env[66641]: INFO nova.compute.manager [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Terminating instance [ 900.225016] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.225016] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.271024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.271024] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquired lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 900.271024] env[66641]: DEBUG nova.network.neutron [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 900.450439] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146246, 'name': PowerOffVM_Task, 'duration_secs': 0.226919} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.450733] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.450890] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.451232] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7bdb889-9c4b-4a84-9df1-619dd5d81d8b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.530633] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.530792] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Deleting contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.530994] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Deleting the datastore file [datastore1] 88ae00a2-6139-4258-b316-0f75032275ec {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.531355] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48f948f6-201c-41ea-a2b8-c6691de90254 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.538642] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for the task: (returnval){ [ 900.538642] env[66641]: value = "task-5146248" [ 900.538642] env[66641]: _type = "Task" [ 900.538642] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.549587] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146248, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.567923] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.568461] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.580329] env[66641]: DEBUG nova.compute.manager [req-a4b73237-5921-41b3-bc55-be980664d4e8 req-8a60167e-9728-459b-be22-c94b9a2bb7ec service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Received event network-vif-deleted-e873e765-2324-4273-81ba-e543ebc3b867 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 900.580627] env[66641]: INFO nova.compute.manager [req-a4b73237-5921-41b3-bc55-be980664d4e8 req-8a60167e-9728-459b-be22-c94b9a2bb7ec service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Neutron deleted interface e873e765-2324-4273-81ba-e543ebc3b867; detaching it from the instance and deleting it from the info cache [ 900.580811] env[66641]: DEBUG nova.network.neutron [req-a4b73237-5921-41b3-bc55-be980664d4e8 req-8a60167e-9728-459b-be22-c94b9a2bb7ec service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 900.617226] env[66641]: DEBUG nova.compute.manager [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 900.617540] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 900.618772] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e7fd53-c6c5-4e0d-a1f6-db18a55325ef {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.629983] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 900.631626] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c14e4eb-53eb-4f58-9cc2-3f81d84f825d {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.640681] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 900.640681] env[66641]: value = "task-5146249" [ 900.640681] env[66641]: _type = "Task" [ 900.640681] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.651978] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146249, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.776462] env[66641]: WARNING openstack [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 900.776864] env[66641]: WARNING openstack [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 900.781775] env[66641]: DEBUG nova.network.neutron [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 900.817684] env[66641]: DEBUG nova.network.neutron [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.051251] env[66641]: DEBUG oslo_vmware.api [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Task: {'id': task-5146248, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.196232} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.051750] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.051750] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Deleted contents of the VM from datastore datastore1 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.051920] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.052012] env[66641]: INFO nova.compute.manager [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Took 1.13 seconds to destroy the instance on the hypervisor. [ 901.052347] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.052479] env[66641]: DEBUG nova.compute.manager [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 901.052651] env[66641]: DEBUG nova.network.neutron [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 901.053110] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.053458] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.087106] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce8a51e3-50d6-44db-9c89-bcebe6989a35 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.098305] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.098305] env[66641]: WARNING openstack [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.113676] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7585dbdc-6413-4a66-83b8-1d339f4e9f43 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.132013] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.132368] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.141645] env[66641]: INFO nova.compute.manager [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task possibly preempted: Conflict updating instance cdfcd9d8-dbf7-4046-8338-42762fc389a1. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} [ 901.156914] env[66641]: ERROR oslo_messaging.rpc.server [None req-cddf52fe-8987-42ac-928c-f0b34a0ac815 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Exception during message handling: nova.exception_Remote.UnexpectedDeletingTaskStateError_Remote: Conflict updating instance cdfcd9d8-dbf7-4046-8338-42762fc389a1. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} [ 901.156914] env[66641]: Traceback (most recent call last): [ 901.156914] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 2386, in _instance_update [ 901.156914] env[66641]: update_on_match(compare, 'uuid', updates) [ 901.156914] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/orm.py", line 52, in update_on_match [ 901.156914] env[66641]: return update_match.update_on_match( [ 901.156914] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/update_match.py", line 194, in update_on_match [ 901.156914] env[66641]: raise NoRowsMatched("Zero rows matched for %d attempts" % attempts) [ 901.156914] env[66641]: oslo_db.sqlalchemy.update_match.NoRowsMatched: Zero rows matched for 3 attempts [ 901.156914] env[66641]: During handling of the above exception, another exception occurred: [ 901.156914] env[66641]: Traceback (most recent call last): [ 901.156914] env[66641]: File "/opt/stack/nova/nova/conductor/manager.py", line 143, in _object_dispatch [ 901.156914] env[66641]: return getattr(target, method)(*args, **kwargs) [ 901.156914] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 901.156914] env[66641]: return fn(self, *args, **kwargs) [ 901.156914] env[66641]: File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 901.156914] env[66641]: old_ref, inst_ref = db.instance_update_and_get_original( [ 901.156914] env[66641]: File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 901.156914] env[66641]: return f(*args, **kwargs) [ 901.156914] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 901.156914] env[66641]: with excutils.save_and_reraise_exception() as ectxt: [ 901.156914] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 901.156914] env[66641]: self.force_reraise() [ 901.157583] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 901.157583] env[66641]: raise self.value [ 901.157583] env[66641]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 901.157583] env[66641]: return f(*args, **kwargs) [ 901.157583] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 901.157583] env[66641]: return f(context, *args, **kwargs) [ 901.157583] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 2303, in instance_update_and_get_original [ 901.157583] env[66641]: return (copy.copy(instance_ref), _instance_update( [ 901.157583] env[66641]: File "/opt/stack/nova/nova/db/main/api.py", line 2445, in _instance_update [ 901.157583] env[66641]: raise exc(**exc_props) [ 901.157583] env[66641]: nova.exception.UnexpectedDeletingTaskStateError: Conflict updating instance cdfcd9d8-dbf7-4046-8338-42762fc389a1. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 901.157583] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 161, in decorated_function [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 156, in decorated_function [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 212, in decorated_function [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 901.158191] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 202, in decorated_function [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7667, in unshelve_instance [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server do_unshelve_instance() [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 415, in inner [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7664, in do_unshelve_instance [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server self._unshelve_instance( [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 7785, in _unshelve_instance [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server instance.save(expected_task_state=task_states.SPAWNING) [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 209, in wrapper [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server updates, result = self.indirection_api.object_action( [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/conductor/rpcapi.py", line 247, in object_action [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server return cctxt.call(context, 'object_action', objinst=objinst, [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 180, in call [ 901.158605] env[66641]: ERROR oslo_messaging.rpc.server result = self.transport._send( [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server return self._driver.send(target, ctxt, message, [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 794, in send [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server return self._send(target, ctxt, message, wait_for_reply, timeout, [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/_drivers/amqpdriver.py", line 786, in _send [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server raise result [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server nova.exception_Remote.UnexpectedDeletingTaskStateError_Remote: Conflict updating instance cdfcd9d8-dbf7-4046-8338-42762fc389a1. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/db/main/api.py", line 2386, in _instance_update [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server update_on_match(compare, 'uuid', updates) [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/orm.py", line 52, in update_on_match [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server return update_match.update_on_match( [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/update_match.py", line 194, in update_on_match [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server raise NoRowsMatched("Zero rows matched for %d attempts" % attempts) [ 901.159012] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server oslo_db.sqlalchemy.update_match.NoRowsMatched: Zero rows matched for 3 attempts [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/conductor/manager.py", line 143, in _object_dispatch [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server return getattr(target, method)(*args, **kwargs) [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_versionedobjects/base.py", line 226, in wrapper [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server return fn(self, *args, **kwargs) [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/objects/instance.py", line 878, in save [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server old_ref, inst_ref = db.instance_update_and_get_original( [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/db/utils.py", line 35, in wrapper [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 144, in wrapper [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception() as ectxt: [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server [ 901.159424] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/api.py", line 142, in wrapper [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/db/main/api.py", line 207, in wrapper [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server return f(context, *args, **kwargs) [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/db/main/api.py", line 2303, in instance_update_and_get_original [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server return (copy.copy(instance_ref), _instance_update( [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/db/main/api.py", line 2445, in _instance_update [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server raise exc(**exc_props) [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server nova.exception.UnexpectedDeletingTaskStateError: Conflict updating instance cdfcd9d8-dbf7-4046-8338-42762fc389a1. Expected: {'task_state': ['spawning']}. Actual: {'task_state': 'deleting'} [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.160043] env[66641]: ERROR oslo_messaging.rpc.server [ 901.163061] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146249, 'name': PowerOffVM_Task, 'duration_secs': 0.235327} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.171804] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 901.172277] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 901.172354] env[66641]: DEBUG nova.compute.manager [req-a4b73237-5921-41b3-bc55-be980664d4e8 req-8a60167e-9728-459b-be22-c94b9a2bb7ec service nova] [instance: c95b481a-7956-410d-971c-7d94911230bb] Detach interface failed, port_id=e873e765-2324-4273-81ba-e543ebc3b867, reason: Instance c95b481a-7956-410d-971c-7d94911230bb could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 901.172732] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1e1a33a-c2e2-4f1c-a8bd-1ff50555dbb7 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.254762] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 901.254953] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 901.255480] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleting the datastore file [datastore2] cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 901.255620] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-134f3e24-b6b7-4cf1-b27d-d8cd8adf0bfb {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.264136] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for the task: (returnval){ [ 901.264136] env[66641]: value = "task-5146251" [ 901.264136] env[66641]: _type = "Task" [ 901.264136] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.277888] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146251, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.287266] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.287568] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 901.321712] env[66641]: INFO nova.compute.manager [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 1.54 seconds to deallocate network for instance. [ 901.396138] env[66641]: DEBUG nova.network.neutron [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.412191] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ea8efc-faa8-49d8-9966-350698870cf6 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.424314] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8686e11-9764-4363-b5c0-d824bd5aaf51 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.464349] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f63760c-7b6c-4df1-b093-61d0c516b39a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.472050] env[66641]: DEBUG nova.network.neutron [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updated VIF entry in instance network info cache for port 049d8c0f-cda0-4cd9-b972-188efe21a140. {{(pid=66641) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3545}} [ 901.472050] env[66641]: DEBUG nova.network.neutron [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [{"id": "049d8c0f-cda0-4cd9-b972-188efe21a140", "address": "fa:16:3e:15:f5:a4", "network": {"id": "28059287-894a-40a1-8385-d1fddec6f422", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-167219310-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3a9466d8e6fa49c59a14110b417664c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "225b6979-9329-403b-91fa-138bd41f6e83", "external-id": "nsx-vlan-transportzone-38", "segmentation_id": 38, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap049d8c0f-cd", "ovs_interfaceid": "049d8c0f-cda0-4cd9-b972-188efe21a140", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.476044] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58ebb70-ba72-4151-996f-509ffe78755a {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.495831] env[66641]: DEBUG nova.compute.provider_tree [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.775802] env[66641]: DEBUG oslo_vmware.api [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Task: {'id': task-5146251, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.184161} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.776708] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.776899] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.777089] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.777251] env[66641]: INFO nova.compute.manager [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Took 1.16 seconds to destroy the instance on the hypervisor. [ 901.777495] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 901.777683] env[66641]: DEBUG nova.compute.manager [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 901.777779] env[66641]: DEBUG nova.network.neutron [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 901.778304] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.778562] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.790154] env[66641]: DEBUG nova.compute.manager [req-286fd470-9587-4625-9745-72b78412bc26 req-48558aae-3228-41db-8e0b-954b377d7ce0 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Received event network-vif-deleted-049d8c0f-cda0-4cd9-b972-188efe21a140 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 901.790154] env[66641]: INFO nova.compute.manager [req-286fd470-9587-4625-9745-72b78412bc26 req-48558aae-3228-41db-8e0b-954b377d7ce0 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Neutron deleted interface 049d8c0f-cda0-4cd9-b972-188efe21a140; detaching it from the instance and deleting it from the info cache [ 901.790154] env[66641]: DEBUG nova.network.neutron [req-286fd470-9587-4625-9745-72b78412bc26 req-48558aae-3228-41db-8e0b-954b377d7ce0 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.826317] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 901.826661] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 901.835491] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 901.899647] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Releasing lock "refresh_cache-a8d805ee-1838-4e22-a2de-a48848b14f9c" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.899759] env[66641]: DEBUG nova.compute.manager [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 901.900622] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 901.900901] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53b735ba-8752-404b-a6ad-811e3823806c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.911260] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Powering off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 901.911260] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6100a81b-ee33-4f37-ba67-8bdf406b6a55 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.918369] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 901.918369] env[66641]: value = "task-5146252" [ 901.918369] env[66641]: _type = "Task" [ 901.918369] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.928692] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146252, 'name': PowerOffVM_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.974719] env[66641]: DEBUG oslo_concurrency.lockutils [req-dfe493cd-b890-47f0-b565-1d8a5d2418a8 req-ab24b242-72b9-46cb-b3ff-6c3aa0fc9c1b service nova] Releasing lock "refresh_cache-88ae00a2-6139-4258-b316-0f75032275ec" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 901.985050] env[66641]: DEBUG nova.network.neutron [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 901.999230] env[66641]: DEBUG nova.scheduler.client.report [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 902.251951] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.292020] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab14ac03-ea27-4a0d-826d-e818c39f9e20 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.301937] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1947091-fb59-4289-8a5b-d5dbf14f031e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.332236] env[66641]: DEBUG nova.compute.manager [req-286fd470-9587-4625-9745-72b78412bc26 req-48558aae-3228-41db-8e0b-954b377d7ce0 service nova] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Detach interface failed, port_id=049d8c0f-cda0-4cd9-b972-188efe21a140, reason: Instance 88ae00a2-6139-4258-b316-0f75032275ec could not be found. {{(pid=66641) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:11669}} [ 902.428902] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146252, 'name': PowerOffVM_Task, 'duration_secs': 0.141745} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.429254] env[66641]: DEBUG nova.virt.vmwareapi.vm_util [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Powered off the VM {{(pid=66641) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.429733] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Unregistering the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.429733] env[66641]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bb1ce5b8-5ecb-4876-aaa4-c6571a52672b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.464265] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Unregistered the VM {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.464709] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Deleting contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.464918] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Deleting the datastore file [datastore2] a8d805ee-1838-4e22-a2de-a48848b14f9c {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.465239] env[66641]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5506043-13f5-4d21-9f5d-2a8455caed22 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.473637] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for the task: (returnval){ [ 902.473637] env[66641]: value = "task-5146254" [ 902.473637] env[66641]: _type = "Task" [ 902.473637] env[66641]: } to complete. {{(pid=66641) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.485664] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146254, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.489219] env[66641]: INFO nova.compute.manager [-] [instance: 88ae00a2-6139-4258-b316-0f75032275ec] Took 1.44 seconds to deallocate network for instance. [ 902.497698] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "beb86b7b-71d1-4790-b464-abe79ff81c1f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 902.497927] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "beb86b7b-71d1-4790-b464-abe79ff81c1f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.504443] env[66641]: DEBUG oslo_concurrency.lockutils [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.217s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 902.504638] env[66641]: INFO nova.compute.manager [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Successfully reverted task state from image_uploading on failure for instance. [ 902.507030] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.671s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 902.507030] env[66641]: DEBUG nova.objects.instance [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lazy-loading 'resources' on Instance uuid c95b481a-7956-410d-971c-7d94911230bb {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server [None req-e65aecb3-960a-4d2d-81f2-eedb25e446e7 tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Exception during message handling: oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created [ 902.509138] env[66641]: Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created' [ 902.509138] env[66641]: Faults: [ManagedObjectNotFound] [ 902.509138] env[66641]: Details: {'obj': 'snapshot-1000732'} [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server response = request(managed_object, **kwargs) [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__ [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server return client.invoke(args, kwargs) [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server result = self.send(soapenv, timeout=timeout) [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server return self.process_reply(reply.message, None, None) [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply [ 902.509138] env[66641]: ERROR oslo_messaging.rpc.server raise WebFault(fault, replyroot) [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server suds.WebFault: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created' [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server return api_method(*args, **kwargs) [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server raise exceptions.VimFaultException(fault_list, fault_string, [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server oslo_vmware.exceptions.VimFaultException: The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created' [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server Faults: [ManagedObjectNotFound] [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server Details: {'obj': 'snapshot-1000732'} [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 174, in _process_incoming [ 902.509515] env[66641]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 165, in decorated_function [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 902.509952] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 156, in decorated_function [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1483, in decorated_function [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 212, in decorated_function [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 202, in decorated_function [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4564, in backup_instance [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server self._do_snapshot_instance(context, image_id, instance) [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 232, in decorated_function [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 902.510654] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 229, in decorated_function [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server return function(self, context, image_id, instance, [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4550, in _do_snapshot_instance [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server self._snapshot_instance(context, image_id, instance, [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 4633, in _snapshot_instance [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server self.driver.snapshot(context, instance, image_id, [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 573, in snapshot [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server self._vmops.snapshot(context, instance, image_id, update_task_state) [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1033, in snapshot [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server self._delete_vm_snapshot(instance, vm_ref, snapshot_ref) [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/decorator.py", line 235, in fun [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server return caller(func, *(extras + args), **kw) [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 124, in retry_if_task_in_progress [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server f(*args, **kwargs) [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 933, in _delete_vm_snapshot [ 902.511134] env[66641]: ERROR oslo_messaging.rpc.server delete_snapshot_task = self._session._call_method( [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 127, in _call_method [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception() as ctxt: [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 256, in __exit__ [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 222, in force_reraise [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server raise self.value [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 125, in _call_method [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server return self.invoke_api(module, method, *args, **kwargs) [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server return _invoke_api(module, method, *args, **kwargs) [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server return evt.wait() [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server result = hub.switch() [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server return self.greenlet.switch() [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner [ 902.511594] env[66641]: ERROR oslo_messaging.rpc.server idle = self.f(*self.args, **self.kw) [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server result = f(*args, **kwargs) [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server raise clazz(str(excep), [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server Cause: Server raised fault: 'The object 'vim.vm.Snapshot:snapshot-1000732' has already been deleted or has not been completely created' [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server Faults: [ManagedObjectNotFound] [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server Details: {'obj': 'snapshot-1000732'} [ 902.511996] env[66641]: ERROR oslo_messaging.rpc.server [ 902.599061] env[66641]: DEBUG nova.network.neutron [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 902.985075] env[66641]: DEBUG oslo_vmware.api [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Task: {'id': task-5146254, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09501} completed successfully. {{(pid=66641) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.985075] env[66641]: DEBUG nova.virt.vmwareapi.ds_util [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Deleted the datastore file {{(pid=66641) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.985075] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Deleted contents of the VM from datastore datastore2 {{(pid=66641) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 902.985075] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 902.985727] env[66641]: INFO nova.compute.manager [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Took 1.09 seconds to destroy the instance on the hypervisor. [ 902.985727] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 902.985841] env[66641]: DEBUG nova.compute.manager [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 902.985889] env[66641]: DEBUG nova.network.neutron [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 902.986422] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 902.986817] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 902.995539] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.000325] env[66641]: DEBUG nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Starting instance... {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2475}} [ 903.008077] env[66641]: DEBUG nova.network.neutron [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 903.008611] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 903.008857] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 903.085900] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cd7f8a9-1719-4993-b587-7f9b6945774c {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.093947] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76f15aa5-6ae1-460e-b86b-4d07f5934eda {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.126477] env[66641]: INFO nova.compute.manager [-] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Took 1.35 seconds to deallocate network for instance. [ 903.129299] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144f30fa-509d-44dd-ac1c-8c974e3db2ed {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.140297] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd66d05-a827-4787-9a50-3d5848ce5e96 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.156973] env[66641]: DEBUG nova.compute.provider_tree [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.517498] env[66641]: DEBUG nova.network.neutron [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 903.519766] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.636800] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 903.660443] env[66641]: DEBUG nova.scheduler.client.report [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 903.816124] env[66641]: DEBUG nova.compute.manager [req-329fac8c-c668-4837-88b7-b8f48621ee56 req-0c484eff-080b-4914-9ad0-5096f24a5321 service nova] [instance: cdfcd9d8-dbf7-4046-8338-42762fc389a1] Received event network-vif-deleted-8bed634b-5119-478e-a71f-a832ef16f035 {{(pid=66641) external_instance_event /opt/stack/nova/nova/compute/manager.py:11835}} [ 904.021074] env[66641]: INFO nova.compute.manager [-] [instance: a8d805ee-1838-4e22-a2de-a48848b14f9c] Took 1.03 seconds to deallocate network for instance. [ 904.165424] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.167793] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.172s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.168088] env[66641]: DEBUG nova.objects.instance [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lazy-loading 'resources' on Instance uuid 88ae00a2-6139-4258-b316-0f75032275ec {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.184928] env[66641]: INFO nova.scheduler.client.report [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Deleted allocations for instance c95b481a-7956-410d-971c-7d94911230bb [ 904.527687] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.692137] env[66641]: DEBUG oslo_concurrency.lockutils [None req-ca565756-ee06-425c-9a02-81d7b4e7e18c tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.587s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.693476] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.443s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.693709] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "c95b481a-7956-410d-971c-7d94911230bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 904.693912] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 904.694099] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 904.698050] env[66641]: INFO nova.compute.manager [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Terminating instance [ 904.715381] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb0b528-c974-45b8-9867-5270b52be5dc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.724607] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d19c192-9d2e-4c69-b31b-db6f54918dee {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.759388] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f999df-1995-445a-b2d9-c1cc134f619f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.768168] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559a1e4a-d10d-4876-bb6a-c5cedeb0d962 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.782271] env[66641]: DEBUG nova.compute.provider_tree [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.201677] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquiring lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.201887] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Acquired lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 905.202076] env[66641]: DEBUG nova.network.neutron [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 905.286188] env[66641]: DEBUG nova.scheduler.client.report [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 905.704957] env[66641]: DEBUG nova.compute.utils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Can not refresh info_cache because instance was not found {{(pid=66641) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 905.705760] env[66641]: WARNING openstack [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 905.706288] env[66641]: WARNING openstack [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 905.711684] env[66641]: DEBUG nova.network.neutron [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 905.791083] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 905.793035] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.273s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 905.794645] env[66641]: INFO nova.compute.claims [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 905.808178] env[66641]: INFO nova.scheduler.client.report [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Deleted allocations for instance 88ae00a2-6139-4258-b316-0f75032275ec [ 905.832374] env[66641]: DEBUG nova.network.neutron [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 906.300442] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 0.506s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.301815] env[66641]: DEBUG nova.compute.utils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Instance beb86b7b-71d1-4790-b464-abe79ff81c1f could not be found. {{(pid=66641) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:431}} [ 906.303944] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.667s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 906.304092] env[66641]: DEBUG nova.objects.instance [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lazy-loading 'resources' on Instance uuid cdfcd9d8-dbf7-4046-8338-42762fc389a1 {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.306509] env[66641]: DEBUG nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Instance disappeared during build. {{(pid=66641) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2556}} [ 906.306680] env[66641]: DEBUG nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Unplugging VIFs for instance {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3050}} [ 906.306896] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquiring lock "refresh_cache-beb86b7b-71d1-4790-b464-abe79ff81c1f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.307065] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Acquired lock "refresh_cache-beb86b7b-71d1-4790-b464-abe79ff81c1f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:316}} [ 906.308729] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Building network info cache for instance {{(pid=66641) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 906.315329] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4c54b415-9690-432d-997f-c8f6590bb246 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "88ae00a2-6139-4258-b316-0f75032275ec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.901s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 906.335539] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Releasing lock "refresh_cache-c95b481a-7956-410d-971c-7d94911230bb" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 906.335975] env[66641]: DEBUG nova.compute.manager [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Start destroying the instance on the hypervisor. {{(pid=66641) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3201}} [ 906.336187] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Destroying instance {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.336473] env[66641]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-55b24332-825f-4e5e-ac80-26d5e0ca58cc {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.347829] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3c1a54f-8ea4-4fa2-a58d-71d3900ca457 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.374305] env[66641]: WARNING nova.virt.vmwareapi.vmops [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c95b481a-7956-410d-971c-7d94911230bb could not be found. [ 906.374509] env[66641]: DEBUG nova.virt.vmwareapi.vmops [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance destroyed {{(pid=66641) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.374687] env[66641]: INFO nova.compute.manager [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 0.04 seconds to destroy the instance on the hypervisor. [ 906.374933] env[66641]: DEBUG oslo.service.backend._common.loopingcall [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=66641) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/backend/_common/loopingcall.py:419}} [ 906.375482] env[66641]: DEBUG nova.compute.manager [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 906.375644] env[66641]: DEBUG nova.network.neutron [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 906.376151] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.376437] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.396865] env[66641]: DEBUG nova.network.neutron [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 906.397377] env[66641]: WARNING openstack [-] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.397706] env[66641]: WARNING openstack [-] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.812768] env[66641]: DEBUG nova.compute.utils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Can not refresh info_cache because instance was not found {{(pid=66641) refresh_info_cache_for_instance /opt/stack/nova/nova/compute/utils.py:1056}} [ 906.813344] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 906.813816] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 906.819063] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 906.839213] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a35b807-0948-4fdd-bda5-3d5c7b87be26 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.847411] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9845c14c-f57f-44d7-8cf9-00943bc2ac66 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.879082] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8e32ae-fde2-4fa2-a052-c6fc010df08e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.889153] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525d8355-1d89-4c7e-b882-82dab5540063 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.903949] env[66641]: DEBUG nova.compute.provider_tree [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 906.905317] env[66641]: DEBUG nova.network.neutron [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 906.925256] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 907.407838] env[66641]: DEBUG nova.scheduler.client.report [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 907.411077] env[66641]: INFO nova.compute.manager [-] [instance: c95b481a-7956-410d-971c-7d94911230bb] Took 1.04 seconds to deallocate network for instance. [ 907.427955] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Releasing lock "refresh_cache-beb86b7b-71d1-4790-b464-abe79ff81c1f" {{(pid=66641) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:334}} [ 907.428254] env[66641]: DEBUG nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=66641) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3073}} [ 907.428435] env[66641]: DEBUG nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Deallocating network for instance {{(pid=66641) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2327}} [ 907.429259] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] deallocate_for_instance() {{(pid=66641) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1857}} [ 907.429911] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.429911] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.452211] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Instance cache missing network info. {{(pid=66641) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3388}} [ 907.453029] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'block-storage': Encountered an exception attempting to process config for project 'cinder' (service type 'block-storage'): no such option valid_interfaces in group [cinder]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [cinder] [ 907.453519] env[66641]: WARNING openstack [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Disabling service 'key-manager': Encountered an exception attempting to process config for project 'barbican' (service type 'key-manager'): no such option valid_interfaces in group [barbican]: oslo_config.cfg.NoSuchOptError: no such option valid_interfaces in group [barbican] [ 907.914646] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.917116] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.390s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 907.917347] env[66641]: DEBUG nova.objects.instance [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lazy-loading 'resources' on Instance uuid a8d805ee-1838-4e22-a2de-a48848b14f9c {{(pid=66641) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 907.918541] env[66641]: INFO nova.compute.manager [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] [instance: c95b481a-7956-410d-971c-7d94911230bb] Instance disappeared during terminate [ 907.918710] env[66641]: DEBUG oslo_concurrency.lockutils [None req-6db51dd5-bb2f-4a7a-aa97-b4766b8a27de tempest-ServerActionsTestOtherB-1735812246 tempest-ServerActionsTestOtherB-1735812246-project-member] Lock "c95b481a-7956-410d-971c-7d94911230bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.225s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 907.948779] env[66641]: INFO nova.scheduler.client.report [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Deleted allocations for instance cdfcd9d8-dbf7-4046-8338-42762fc389a1 [ 907.960909] env[66641]: DEBUG nova.network.neutron [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Updating instance_info_cache with network_info: [] {{(pid=66641) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:121}} [ 908.435308] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52bf2039-6190-4469-bf61-1372716fb21f {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.444191] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20f59dea-af88-4e19-9985-545362dd15b9 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.478424] env[66641]: INFO nova.compute.manager [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] [instance: beb86b7b-71d1-4790-b464-abe79ff81c1f] Took 1.05 seconds to deallocate network for instance. [ 908.481217] env[66641]: DEBUG oslo_concurrency.lockutils [None req-4b8e4f4a-fb1d-4a93-bcf0-355be3204093 tempest-AttachVolumeShelveTestJSON-848866166 tempest-AttachVolumeShelveTestJSON-848866166-project-member] Lock "cdfcd9d8-dbf7-4046-8338-42762fc389a1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.374s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 908.482621] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df8dc3d8-8e71-4da9-aa73-135c64d18f31 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.492774] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc8d7e7-4531-4f4a-97d3-76e8e277c34e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.510459] env[66641]: DEBUG nova.compute.provider_tree [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.015487] env[66641]: DEBUG nova.scheduler.client.report [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 909.501656] env[66641]: DEBUG oslo_concurrency.lockutils [None req-7f29642f-8001-46da-a89e-91171bac40b6 tempest-AttachVolumeNegativeTest-1310596129 tempest-AttachVolumeNegativeTest-1310596129-project-member] Lock "beb86b7b-71d1-4790-b464-abe79ff81c1f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 7.004s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 909.518973] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 910.039766] env[66641]: DEBUG oslo_concurrency.lockutils [None req-1128b91b-36d8-406d-b2de-08cd0671ba03 tempest-ServerShowV257Test-673484002 tempest-ServerShowV257Test-673484002-project-member] Lock "a8d805ee-1838-4e22-a2de-a48848b14f9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.284s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 912.932491] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.436927] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.437147] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.437308] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 914.184239] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 914.184689] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager.update_available_resource {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 914.688354] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.688605] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 914.688789] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 914.688979] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=66641) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:937}} [ 914.689957] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d549dda6-beb1-4f3a-9a96-98b872879164 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.698960] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-364f7f70-3f3c-473e-aaad-ce80794e2661 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.712526] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4116e0-05c1-479b-88b3-7aaceea2cc17 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.719297] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc46976-e694-4cf6-908f-9220af2f563b {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.748794] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179895MB free_disk=95GB free_vcpus=48 pci_devices=None {{(pid=66641) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1136}} [ 914.748982] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:405}} [ 914.749166] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:410}} [ 915.770226] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1159}} [ 915.770585] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=100GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] stats={'failed_builds': '2'} {{(pid=66641) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1168}} [ 915.783743] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-438f62eb-fafa-4074-a076-b5eb8f9594e0 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.791812] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e3dae7b-7964-4426-8598-6b74dcba810e {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.822183] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9f2be4-8bc8-476b-a3de-41877b1f8d52 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.830533] env[66641]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-064762a1-1882-49d1-b358-cfd8895a6796 {{(pid=66641) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.845127] env[66641]: DEBUG nova.compute.provider_tree [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed in ProviderTree for provider: 750ffd2d-5e46-4240-a614-995f2be7c9cb {{(pid=66641) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.348673] env[66641]: DEBUG nova.scheduler.client.report [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Inventory has not changed for provider 750ffd2d-5e46-4240-a614-995f2be7c9cb based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 200, 'reserved': 0, 'min_unit': 1, 'max_unit': 95, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=66641) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:958}} [ 916.853520] env[66641]: DEBUG nova.compute.resource_tracker [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=66641) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1097}} [ 916.853916] env[66641]: DEBUG oslo_concurrency.lockutils [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.105s {{(pid=66641) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:424}} [ 918.854058] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.854058] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.854058] env[66641]: DEBUG oslo_service.periodic_task [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=66641) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.854449] env[66641]: DEBUG nova.compute.manager [None req-045e5f88-19e8-443a-a6e0-94e655c7b707 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=66641) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:11251}}